Giter Site home page Giter Site logo

dinghanshen / swem Goto Github PK

View Code? Open in Web Editor NEW
283.0 283.0 54.0 7.92 MB

The Tensorflow code for this ACL 2018 paper: "Baseline Needs More Love: On Simple Word-Embedding-Based Models and Associated Pooling Mechanisms"

Python 65.83% Makefile 0.17% C++ 1.52% Jupyter Notebook 32.48%
deep-learning natural-language-processing representation-learning tensorflow

swem's People

Contributors

chunyuanli avatar dinghanshen avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

swem's Issues

Difference between the paper and the implementation

Thanks sharing us the code.
I have question regarding DBPedia experiment code.

https://github.com/dinghanshen/SWEM/blob/master/eval_dbpedia_emb.py#L96-L118

It seems to me that you are applying max-pooling (tf.nn.max_pool) then another max pooling while your paper says that you applied average-pooling then max pooling.
Also, pooling window size (max_len option) seems to be very large, it seems to me like you are just applying max pooling rather than hierarchical pooling.

Is this intended implementation? Or am I missing somthing?

Not find SWEM-hier

Hi, seems not find hier encoder as paper mentioned. Very interested to see it :)

aver_emb_encoder

I don't think aver_emb_encoder is actually ever defined in the committed code base--though it's pretty straightforward to understand what it should be doing :-).

Package dependency related errors occuring while loading snli_emb.py

While loading the snli_emb.py with current numpy==1.16.* the following error came:

  File "/usr/local/lib/python2.7/dist-packages/numpy/lib/npyio.py", line 451, in load
    raise ValueError("Cannot load file containing pickled data "
ValueError: Cannot load file containing pickled data when allow_pickle=False

When, I downgraded it to numpy==1.14.* the error related to binary compatibility came:

ValueError: numpy.ufunc size changed, may indicate binary incompatibility. Expected 216 from C header, got 192 from PyObject

To me it looks like a package version compatibility related issue with respect to numpy used for creating snli_emb.p pickle file.

It'll be really helpful if you can mention the package versions also that you have used in as your import statements like for gensim, scipy etc packages.

Experiment Setting on Section 4.1.1 in the paper

It seems that the repo do not contain code about the section 4.1.1: Interpreting model predictions, but I am really interested in the part, could you share it if you have time.

Also, I have some questions on the experiment setting of the section.
(1) Do you set the word embedding randomly initialized(as you mention in the section) but not from Glove pretrain.
(2) What is your network architecture, Is Word_embedding + MaxPooling +Classifier or Word_Embedding + MaxPooling + MLP + Classifier.

About NASH

 I am a first-year doctoral student at Beijing University of Posts and Telecommunications,China, and I major in computer science and technology.I have recently read your paper"NASH: Toward End-to-End Neural Architecture for Generative Semantic Hashing" .It would make a really positive contribution to my work. I am wondering if you could kindly send me the source program of your experiments. I promise they will be used only for research purposed.

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.