Giter Site home page Giter Site logo

ropensci / elastic Goto Github PK

View Code? Open in Web Editor NEW
246.0 27.0 58.0 3.33 MB

R client for the Elasticsearch HTTP API

Home Page: https://docs.ropensci.org/elastic

License: Other

Makefile 0.24% R 99.76%
elasticsearch http database database-wrapper json rstats r r-package data-science etl

elastic's Introduction

elastic

Project Status: Active – The project has reached a stable, usable state and is being actively developed. R-check cran checks rstudio mirror downloads cran version

A general purpose R interface to Elasticsearch

Elasticsearch info

Compatibility

This client is developed following the latest stable releases, currently v7.10.0. It is generally compatible with older versions of Elasticsearch. Unlike the Python client, we try to keep as much compatibility as possible within a single version of this client, as that's an easier setup in R world.

Security

You're fine running ES locally on your machine, but be careful just throwing up ES on a server with a public IP address - make sure to think about security.

  • Elastic has paid products - but probably only applicable to enterprise users
  • DIY security - there are a variety of techniques for securing your Elasticsearch installation. A number of resources are collected in a blog post - tools include putting your ES behind something like Nginx, putting basic auth on top of it, using https, etc.

Installation

Stable version from CRAN

install.packages("elastic")

Development version from GitHub

remotes::install_github("ropensci/elastic")
library('elastic')

Install Elasticsearch

w/ Docker

Pull the official elasticsearch image

# elasticsearch needs to have a version tag. We're pulling 7.10.1 here
docker pull elasticsearch:7.10.1

Then start up a container

docker run -d -p 9200:9200 elasticsearch:7.10.1

Then elasticsearch should be available on port 9200, try curl localhost:9200 and you should get the familiar message indicating ES is on.

If you're using boot2docker, you'll need to use the IP address in place of localhost. Get it by doing boot2docker ip.

on OSX

  • Download zip or tar file from Elasticsearch see here for download, e.g., curl -L -O https://artifacts.elastic.co/downloads/elasticsearch/elasticsearch-7.10.0-darwin-x86_64.tar.gz
  • Extract: tar -zxvf elasticsearch-7.10.0-darwin-x86_64.tar.gz
  • Move it: sudo mv elasticsearch-7.10.0 /usr/local
  • Navigate to /usr/local: cd /usr/local
  • Delete symlinked elasticsearch directory: rm -rf elasticsearch
  • Add shortcut: sudo ln -s elasticsearch-7.10.0 elasticsearch (replace version with your version)

You can also install via Homebrew: brew install elasticsearch

Note: for the 1.6 and greater upgrades of Elasticsearch, they want you to have java 8 or greater. I downloaded Java 8 from here http://www.oracle.com/technetwork/java/javase/downloads/jdk8-downloads-2133151.html and it seemed to work great.

Upgrading Elasticsearch

I am not totally clear on best practice here, but from what I understand, when you upgrade to a new version of Elasticsearch, place old elasticsearch/data and elasticsearch/config directories into the new installation (elasticsearch/ dir). The new elasticsearch instance with replaced data and config directories should automatically update data to the new version and start working. Maybe if you use homebrew on a Mac to upgrade it takes care of this for you - not sure.

Obviously, upgrading Elasticsearch while keeping it running is a different thing (some help here from Elastic).

Start Elasticsearch

  • Navigate to elasticsearch: cd /usr/local/elasticsearch
  • Start elasticsearch: bin/elasticsearch

I create a little bash shortcut called es that does both of the above commands in one step (cd /usr/local/elasticsearch && bin/elasticsearch).

Initialization

The function connect() is used before doing anything else to set the connection details to your remote or local elasticsearch store. The details created by connect() are written to your options for the current session, and are used by elastic functions.

x <- connect(port = 9200)

If you're following along here with a local instance of Elasticsearch, you'll use x below to do more stuff.

For AWS hosted elasticsearch, make sure to specify path = "" and the correct port - transport schema pair.

connect(host = <aws_es_endpoint>, path = "", port = 80, transport_schema  = "http")
  # or
connect(host = <aws_es_endpoint>, path = "", port = 443, transport_schema  = "https")

If you are using Elastic Cloud or an installation with authentication (X-pack), make sure to specify path = "", user = "", pwd = "" and the correct port - transport schema pair.

connect(host = <ec_endpoint>, path = "", user="test", pwd = "1234", port = 9243, transport_schema  = "https")

Get some data

Elasticsearch has a bulk load API to load data in fast. The format is pretty weird though. It's sort of JSON, but would pass no JSON linter. I include a few data sets in elastic so it's easy to get up and running, and so when you run examples in this package they'll actually run the same way (hopefully).

I have prepare a non-exported function useful for preparing the weird format that Elasticsearch wants for bulk data loads, that is somewhat specific to PLOS data (See below), but you could modify for your purposes. See make_bulk_plos() and make_bulk_gbif() here.

Shakespeare data

Elasticsearch provides some data on Shakespeare plays. I've provided a subset of this data in this package. Get the path for the file specific to your machine:

shakespeare <- system.file("examples", "shakespeare_data.json", package = "elastic")
# If you're on Elastic v6 or greater, use this one
shakespeare <- system.file("examples", "shakespeare_data_.json", package = "elastic")
shakespeare <- type_remover(shakespeare)

Then load the data into Elasticsearch:

make sure to create your connection object with connect()

# x <- connect()  # do this now if you didn't do this above
invisible(docs_bulk(x, shakespeare))

If you need some big data to play with, the shakespeare dataset is a good one to start with. You can get the whole thing and pop it into Elasticsearch (beware, may take up to 10 minutes or so.):

curl -XGET https://download.elastic.co/demos/kibana/gettingstarted/shakespeare_6.0.json > shakespeare.json
curl -XPUT localhost:9200/_bulk --data-binary @shakespeare.json

Public Library of Science (PLOS) data

A dataset inluded in the elastic package is metadata for PLOS scholarly articles. Get the file path, then load:

if (index_exists(x, "plos")) index_delete(x, "plos")
plosdat <- system.file("examples", "plos_data.json", package = "elastic")
plosdat <- type_remover(plosdat)
invisible(docs_bulk(x, plosdat))

Global Biodiversity Information Facility (GBIF) data

A dataset inluded in the elastic package is data for GBIF species occurrence records. Get the file path, then load:

if (index_exists(x, "gbif")) index_delete(x, "gbif")
gbifdat <- system.file("examples", "gbif_data.json", package = "elastic")
gbifdat <- type_remover(gbifdat)
invisible(docs_bulk(x, gbifdat))

GBIF geo data with a coordinates element to allow geo_shape queries

if (index_exists(x, "gbifgeo")) index_delete(x, "gbifgeo")
gbifgeo <- system.file("examples", "gbif_geo.json", package = "elastic")
gbifgeo <- type_remover(gbifgeo)
invisible(docs_bulk(x, gbifgeo))

More data sets

There are more datasets formatted for bulk loading in the sckott/elastic_data GitHub repository. Find it at https://github.com/sckott/elastic_data


Search

Search the plos index and only return 1 result

Search(x, index = "plos", size = 1)$hits$hits
#> [[1]]
#> [[1]]$`_index`
#> [1] "plos"
#> 
#> [[1]]$`_type`
#> [1] "_doc"
#> 
#> [[1]]$`_id`
#> [1] "0"
#> 
#> [[1]]$`_score`
#> [1] 1
#> 
#> [[1]]$`_source`
#> [[1]]$`_source`$id
#> [1] "10.1371/journal.pone.0007737"
#> 
#> [[1]]$`_source`$title
#> [1] "Phospholipase C-\u03b24 Is Essential for the Progression of the Normal Sleep Sequence and Ultradian Body Temperature Rhythms in Mice"

Search the plos index, and query for antibody, limit to 1 result

Search(x, index = "plos", q = "antibody", size = 1)$hits$hits
#> [[1]]
#> [[1]]$`_index`
#> [1] "plos"
#> 
#> [[1]]$`_type`
#> [1] "_doc"
#> 
#> [[1]]$`_id`
#> [1] "813"
#> 
#> [[1]]$`_score`
#> [1] 5.18676
#> 
#> [[1]]$`_source`
#> [[1]]$`_source`$id
#> [1] "10.1371/journal.pone.0107638"
#> 
#> [[1]]$`_source`$title
#> [1] "Sortase A Induces Th17-Mediated and Antibody-Independent Immunity to Heterologous Serotypes of Group A Streptococci"

Get documents

Get document with id=4

docs_get(x, index = 'plos', id = 4)
#> $`_index`
#> [1] "plos"
#> 
#> $`_type`
#> [1] "_doc"
#> 
#> $`_id`
#> [1] "4"
#> 
#> $`_version`
#> [1] 1
#> 
#> $`_seq_no`
#> [1] 4
#> 
#> $`_primary_term`
#> [1] 1
#> 
#> $found
#> [1] TRUE
#> 
#> $`_source`
#> $`_source`$id
#> [1] "10.1371/journal.pone.0107758"
#> 
#> $`_source`$title
#> [1] "Lactobacilli Inactivate Chlamydia trachomatis through Lactic Acid but Not H2O2"

Get certain fields

docs_get(x, index = 'plos', id = 4, fields = 'id')
#> $`_index`
#> [1] "plos"
#> 
#> $`_type`
#> [1] "_doc"
#> 
#> $`_id`
#> [1] "4"
#> 
#> $`_version`
#> [1] 1
#> 
#> $`_seq_no`
#> [1] 4
#> 
#> $`_primary_term`
#> [1] 1
#> 
#> $found
#> [1] TRUE

Get multiple documents via the multiget API

Same index and different document ids

docs_mget(x, index = "plos", id = 1:2)
#> $docs
#> $docs[[1]]
#> $docs[[1]]$`_index`
#> [1] "plos"
#> 
#> $docs[[1]]$`_type`
#> [1] "_doc"
#> 
#> $docs[[1]]$`_id`
#> [1] "1"
#> 
#> $docs[[1]]$`_version`
#> [1] 1
#> 
#> $docs[[1]]$`_seq_no`
#> [1] 1
#> 
#> $docs[[1]]$`_primary_term`
#> [1] 1
#> 
#> $docs[[1]]$found
#> [1] TRUE
#> 
#> $docs[[1]]$`_source`
#> $docs[[1]]$`_source`$id
#> [1] "10.1371/journal.pone.0098602"
#> 
#> $docs[[1]]$`_source`$title
#> [1] "Population Genetic Structure of a Sandstone Specialist and a Generalist Heath Species at Two Levels of Sandstone Patchiness across the Strait of Gibraltar"
#> 
#> 
#> 
#> $docs[[2]]
#> $docs[[2]]$`_index`
#> [1] "plos"
#> 
#> $docs[[2]]$`_type`
#> [1] "_doc"
#> 
#> $docs[[2]]$`_id`
#> [1] "2"
#> 
#> $docs[[2]]$`_version`
#> [1] 1
#> 
#> $docs[[2]]$`_seq_no`
#> [1] 2
#> 
#> $docs[[2]]$`_primary_term`
#> [1] 1
#> 
#> $docs[[2]]$found
#> [1] TRUE
#> 
#> $docs[[2]]$`_source`
#> $docs[[2]]$`_source`$id
#> [1] "10.1371/journal.pone.0107757"
#> 
#> $docs[[2]]$`_source`$title
#> [1] "Cigarette Smoke Extract Induces a Phenotypic Shift in Epithelial Cells; Involvement of HIF1\u03b1 in Mesenchymal Transition"

Parsing

You can optionally get back raw json from Search(), docs_get(), and docs_mget() setting parameter raw=TRUE.

For example:

(out <- docs_mget(x, index = "plos", id = 1:2, raw = TRUE))
#> [1] "{\"docs\":[{\"_index\":\"plos\",\"_type\":\"_doc\",\"_id\":\"1\",\"_version\":1,\"_seq_no\":1,\"_primary_term\":1,\"found\":true,\"_source\":{\"id\":\"10.1371/journal.pone.0098602\",\"title\":\"Population Genetic Structure of a Sandstone Specialist and a Generalist Heath Species at Two Levels of Sandstone Patchiness across the Strait of Gibraltar\"}},{\"_index\":\"plos\",\"_type\":\"_doc\",\"_id\":\"2\",\"_version\":1,\"_seq_no\":2,\"_primary_term\":1,\"found\":true,\"_source\":{\"id\":\"10.1371/journal.pone.0107757\",\"title\":\"Cigarette Smoke Extract Induces a Phenotypic Shift in Epithelial Cells; Involvement of HIF1\u03b1 in Mesenchymal Transition\"}}]}"
#> attr(,"class")
#> [1] "elastic_mget"

Then parse

jsonlite::fromJSON(out)
#> $docs
#>   _index _type _id _version _seq_no _primary_term found
#> 1   plos  _doc   1        1       1             1  TRUE
#> 2   plos  _doc   2        1       2             1  TRUE
#>                     _source.id
#> 1 10.1371/journal.pone.0098602
#> 2 10.1371/journal.pone.0107757
#>                                                                                                                                                _source.title
#> 1 Population Genetic Structure of a Sandstone Specialist and a Generalist Heath Species at Two Levels of Sandstone Patchiness across the Strait of Gibraltar
#> 2                                Cigarette Smoke Extract Induces a Phenotypic Shift in Epithelial Cells; Involvement of HIF1\u03b1 in Mesenchymal Transition

Known pain points

  • On secure Elasticsearch servers:
    • HEAD requests don't seem to work, not sure why
    • If you allow only GET requests, a number of functions that require POST requests obviously then won't work. A big one is Search(), but you can use Search_uri() to get around this, which uses GET instead of POST, but you can't pass a more complicated query via the body

Screencast

A screencast introducing the package: vimeo.com/124659179

Meta

  • Please report any issues or bugs
  • License: MIT
  • Get citation information for elastic in R doing citation(package = 'elastic')
  • Please note that this package is released with a Contributor Code of Conduct. By contributing to this project, you agree to abide by its terms.

elastic's People

Contributors

colinfay avatar cphaarmeyer avatar dpmccabe avatar everdark avatar featherduino avatar jeroen avatar lchiffon avatar m-yd avatar maelle avatar nelsonschwarz avatar orenov avatar pieterprovoost avatar rfsaldanha avatar sckott avatar statwonk avatar stevenysw avatar ugosan avatar wsams avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

elastic's Issues

Connect to remote server instead of localhost

es_connect("http://search.r-pkg.org", 9200)

fails with

Error in sprintf("\n  Failed to connect to %s\n  Remember to start Elasticsearch before connecting",  : 
  object 'base' not found

because there is

    if (grepl("localhost|127.0.0.1", auth$base)) 
        base <- paste(auth$base, auth$port, sep = ":")
    res <- tryCatch(GET(base, ...), error = function(e) e)
    if ("error" %in% class(res)) {
        stop(sprintf("\n  Failed to connect to %s\n  Remember to start Elasticsearch before connecting", 
            base), call. = FALSE)
    }

in es_connect.

Is this intentional that one can only connect to localhost?

Unhandy termination when server failure response contains no error description

Cf error_parser function + search_POST/scroll_POST ...
In case of error retruned from the server (i.e. tt$status_code > 202), and in specific circumstances,
if might happen that content(tt) has no $error member.

Hence the first test in error_parser fails like this:

Error in if (grepl("SearchParseException", y)) { : 
  argument is of length zero

-> Would be nice to fail more gracefully.

It happened to me after retrieving 3M+ records from my ES server.
(A quite extreme condition I must admit, memory was almost full, might be the root cause of the problem)
(=> would be hard to reproduce, also.)

query argument not working

The argument q= has no effect in the results. This is the result from querying 'JeSuisCharlie':

The results never match the query criteria. Check this out:

Search(index="firehose", q="#JeSuisCharlie", size=2)$hits$hits -> test2; test2[[2]]$_source$text
[1] "RT @UnDislexico: —Tengo VIH y necesito confesarme.\n—Lo siento, no hay cura."

However, the JSON returned by the server (firehose/_search?q=JeSuisCharlie) works perfectly:

_index: "firehose",
_type: "status",
_id: "555434859586650113",
_score: 2.173656,
_source: {
text: "#IS: #JeSuisCharlie #AlQuaida: non! #JeSuisCharlie! #musilmans: nous sommes tous Charlie!",
created_at: "2015-01-14T18:42:44.000Z",
source: "<a href="http://twitter.com/#!/download/ipad" rel="nofollow">Twitter for iPad</a>",

Am I missing something or there's a problem with the function Search when parsing the argument "q"?

scroll does not seem to handle long scroll_id properly

I have lots of shards, and that results in very long scroll_id's (e.g. 6961 characters long).
As per the documentation of elasticsearch, these should be sent in the body of the request.

It works perfectly with curl, but not in R:

> results = Search(index=index, type=type, q=query, fields=fields, scroll="1m", search_type = "scan")
> scroll_id = results$`_scroll_id`
> results = scroll(scroll_id=scroll_id, scroll="1m")
Error in function (type, msg, asError = TRUE)  : Empty reply from server

Using the body also does not work:

> results = scroll(body=list(scroll_id), scroll="1m")
Error in search_POST(path = "_search/scroll", args = list(scroll = scroll,  : 
  formal argument "body" matched by multiple actual arguments

Elasticsearch DSL?

Not sure if this is possible, but something along the lines of, e.g.

To attempt to reproduce this

'{
  "query": {
    "bool" : {
      "must_not" : {
        "range" : {
          "speech_number" : {
            "from" : 1, "to": 5
           }
         }
       }
     }
  }
}'

perhaps

range( speech_number != c(1,5) )

Or

'{
   "query": {
       "multi_match" : {
           "query" : "henry", 
           "fields":  ["text_entry","play_name"] 
        }
     }
}'

from

query(  "henry" ) %>%
    select(text_entry, play_name)

So, a series of verbs-ish to construct elasticsearch Query DSL queries

es_search for aggregations

hello, thx for this nice interface ES | R
i was wondering if it was possible to use the es_search ( in rstudio) the same way i submit search with curl ( see below a search with curl returning aggregations from 1 index )
i imagine something like es_search(index="myindex", type="mytype", .....) but i do not see how to fill the ...

curl -XPOST  "http://localhost:9200/myindex/mytype/_search" -d '
{
   "size": 0, "aggs": {"stats": {
         "terms": {
            "field": "client_ip"
         } }  }
}

best
philippe

index is ignored in es_get

library(elastic)
# Warning message:
# replacing previous import by ‘rjson::toJSON’ when loading ‘elastic’ 
con <- es_connect("http://search.r-pkg.org", 9200)
es_get(index="cran-devel", type="package", id="igraph")
# Error in es_get(index = "cran-devel", type = "package", id = "igraph") : 
#   client error: (403) Forbidden

The url should be base/index/type/id, AFAIK: http://www.elasticsearch.org/guide/en/elasticsearch/reference/current/docs-get.html

library(httr)
GET("http://search.r-pkg.org:9200/cran-devel/package/igraph")
# Response [http://search.r-pkg.org:9200/cran-devel/package/igraph]
#   Status: 200
#   Content-type: application/json; charset=UTF-8
# {"_index":"cran-devel","_type":"package","_id":"igraph","_version":1,"found":true,
# "_source" : { "Package" : "igraph", "Version" : "0.7.1", 
# ...

JSON Data type of property can not be fetched/queried?

Hi ROpensci,

Just wondering why I can't get the type of my fields? I am expecting the result would be numeric, ie. integer or float.

Using the mapping command as below.

$out <- index_get(index='gasen')
$names(out$gasen$mappings$com$properties$SCORE)
[1] "type"

When executing to get the data type, the result becomes NULL.

$names(out$gasen$mappings$com$properties$SCORE$type)
NULL

Refer to source JSON below.

Is this the default/normal/correct behavior?
Any advise/work-around please.

Sincerely,

Jaime Jr

Source JSON: http://localhost:9200/_plugin/head/:

"state": "open",
"settings": {
    "index": {
        "number_of_replicas": "1",
        "number_of_shards": "2",
        "uuid": "65tNPda7Qe--4qdMjAq74w",
        "version": {
            "created": "1030199"
        }
    }
},

"mappings": {

   "com": {
    "properties": {
        "CHANNEL": {
            "type": "string"
        },
        "SCORE": {
            "type": "float"
        },
        "CONTENT": {
            "type": "string"
        },
        "CREATE ON": {
            "format": "dateOptionalTime",
            "type": "date"
        },
        "ID": {
            "type": "string"
        },
        "SENTIMENT": {
            "type": "string"
        },
        "CREATE BY": {
            "type": "string"
        }
    }
}

},

How to search a range of documents according to the values in a column?

Hi, I am using this elastic package to read data from ES. I have already know how to search according to a certain value, but now I wonder, if there is any founction to query a range of documents according to the values in a column. For example, I have a field named "time", which is like a timestamp (e.g 12345, 12346, 12347,12348, ...) . If I want to get the docements with the field "time" value bertween 12345 and 12347, how can I do that?
(I am not sure if here is the right place for questions. If I post on the wrong place, please just delete it and sorry for the trouble.)

Roadmap

A roadmap of things to do/questions/etc.

  • Right now, there are a few functions to interface with the CouchDB river plugin. Do we want to make interfaces to other plugins? Which are highest priority?
  • Authentication. What's the best approach to handle the various authentication schemes?
  • APIs
  • Downstream methods?
    • Makes sense to give back raw data in json format unless otherwise specified, right?
    • Parse raw data to data.frame, list, etc.

Elasticsearch 1.2.1

Dear Scott,

I have to access an elastic search 1.2.1 data base, but it seems like that it is not supported with elastic, isn't it? When I try to connect like

connect(es_base = "ubicity.ait.ac.at:8080/es/", es_user = NULL, es_pwd = NULL, es_key = NULL, force = FALSE)
(I tried it in several different versions)

I only get as return an empty error prompt
Error: (but no error message)

Is it because of ES 1.2.1 or because I screw it up?

Thx,

Tobias

es_search hits limited?

Hello,
When I use es_search, it tells me results$hits$total is in the 10,000's but then the length of results$hits$hits is only 10. Is this the intended functionality, and if so how do I get all the results?
Thanks

Search fails when using (big) numeric arguments

If the from argument is a numeric type and is very big (e.g. 100'000), the call to Search fails.

The reason is that the argument will be automatically converted to a string by httr (build_url function uses paste0 which uses as.character).
But:

> as.character(100000)
[1] "1e+05"

and ES does not expect this input.

My workaround is to use from=sprintf("%d", current) instead of from=current. (=> not blocking)
This is not expected because the documentation for Search describes implicitly from as a number.

(Note that my workaround does not work if current is already a string.)

I assume the same kind of issue could occur with size, though it's less likely to have big numbers for size.

No result returned by Search if type contains special characters

> type_rp="rp_access+comp_name+protocol"
> ...
> results = Search(index=index, type=type, ...)
(empty results)

Search should URL-encode type:

> type_rp="rp_access%2Bcomp_name%2Bprotocol"
> ...
> results = Search(index=index, type=type, ...)
(14335 hits)

Install under Windows

Hello.

Try to install under windows, gives this error:

  • installing source package 'elastic' ...
    ** R
    ** preparing package for lazy loading
    Warning: replacing previous import by 'rjson::toJSON' when loading 'elastic'
    ** help
    Error in iconv(lines, encoding, "UTF-8", sub = "byte") :
    embedded nul in string: '\item{allow_no_indices}{b\0\023 Whether to ignore if a wildcard indices expression resolves into no'
    ERROR: installing Rd objects failed for package 'elastic'
  • removing 'C:/R/R-3.1.1/library/elastic'

Help me :)

Figure out problem with RJSONIO and rjson

For some reason toJSON from RJSONIO and rjson each only work in some cases so that each have to be imported, which is not ideal, and leads to an error on loading the library. See if something can be done. And ripley will get mad about using ::

docs_mget: Error in status_code(x) : object 'out' not found

Don't know how to debug this error returned by the elasticsearch library:

> library('elastic')
> connect()
uri:       http://127.0.0.1 
port:      9200 
username:  NULL 
password:  NULL 
api key:   NULL 
elasticsearch details:   
      status:                  200 
      name:                    Crimson Dynamo 
      Elasticsearch version:   1.2.1 
      ES version timestamp:    2014-06-03T15:02:52Z 
      lucene version:          4.8
> index="logstash-XXXX-*"
> type="my_type"
> query="service:yyyy"
> count(index=index, type=type, q=query)
[1] 510058
> results = docs_mget(index=index, type=type, fields=c("timestamp", "duration_ms", "service"), q=query)
Error in status_code(x) : object 'out' not found
  • Windows 7 64b
  • R engine 3.1.2, RStudio 0.98.1091, elastic library 0.3.0
  • elasticsearch 1.2.2

docs_get: length(url) == 1 is not TRUE

Don't know how to debug this error returned by the elasticsearch library:

> library('elastic')
> connect()
uri:       http://127.0.0.1 
port:      9200 
username:  NULL 
password:  NULL 
api key:   NULL 
elasticsearch details:   
      status:                  200 
      name:                    Crimson Dynamo 
      Elasticsearch version:   1.2.1 
      ES version timestamp:    2014-06-03T15:02:52Z 
      lucene version:          4.8
> index="logstash-XXXX-*"
> type="my_type"
> query="service:yyyy"
> count(index=index, type=type, q=query)
[1] 510058
> results = docs_get(index=index, type=type, fields=c("timestamp", "duration_ms", "service"), q=query, verbose=TRUE)
Error: length(url) == 1 is not TRUE
  • Windows 7 64b
  • R engine 3.1.2, RStudio 0.98.1091, elastic library 0.3.0
  • elasticsearch 1.2.2

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.