Elasticsearch: Index 2GB of documents at once -


i use the mapper-attachments-plugin index , analyze pdfs in elasticsearch index (1 node, 5 shards). works fine problem initial import of 1.800 pdfs (about 2gb) blob table in mysql database.

i use native java client (transport mode), documents database, encode them base64 , send them elasticsearch. run outofmemory errors after 300 documents. when give more memory import application, elasticsearch master node blocked after.

is there better way import documents? perhaps "bulk-index-client" or so?


Comments

Popular posts from this blog

sublimetext3 - what keyboard shortcut is to comment/uncomment for this script tag in sublime -

java - No use of nillable="0" in SOAP Webservice -

ubuntu - Laravel 5.2 quickstart guide gives Not Found Error -