Giter VIP home page Giter VIP logo

crosslingualcontextualemb's People

Contributors

akkikiki avatar talschuster avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

crosslingualcontextualemb's Issues

section 3.3 in paper

Not a code issue, but I have a question regarding your paper.
For footnote 3 in section3.3 (Supervised Anchored Alignment), it says "In practice, we may have multiple target words for a single source word, and the extension is straight-forward.".

My understanding is, dictionary D is a dictionary for source and target word pairs. However for a single source word that may have multiple target words, denote this as 'i', (e.g bear in source(english) can be translated to multiple target words), D would contain more than one entries of (i,D(i)) . So dictionary D would look like:
....
(source, target)
(bear, target word for 'animal bear')
(bear, target word for 'to have')
....
....

Say,
a = anchor vector for source word bear
b = anchor vector for target word 'animal bear'
c = anchor vector for target word 'to have'

then,
we learn a mapping between (a,b) and (a,c).
Is my understanding correct?

The problem is, it is hard to intuitively understand how this method can lead to results demonstrated in Figure2, where distinct point clouds are aligned with their corresponding distinct words in target

Code for generating embedding visualization

Thank you for the nice work. I read your paper and saw a few embedding visualizations that look very nice, but I can't seem to find the code for generating those in the repo here.

I'm also trying to visualize contextualized embeddings (for BERT). How did you manage to do that? Thank you so much for your help.

How to generate the mapping matrixs for the ELMo of my own language.

I've trained several ELMo weights for other languages (e.g., Finnish, Chinese ...)(I consider to enrich your repo), and now I wanna align them (including each LSTM layer) to English space, as just you did.

But, it seems you did not release the codes for generating such aligning matrix (like such below).
ma

P.S.: note that you only released the code for generating the anchors, and I believe it is nothing to do with the aligning matrix.
Or, if I misunderstand the approach, please show the hints, correct my wrong.

So, may I have your prompt reply concerning this issue?
Thx a lot.

option262.json

Hi, when running demo.py models/option262.json is not present by default, are we supposed to download?

Best alignment mapping for English

Hi,

Thank you for this interesting work. I am currently working on extending this approach on top of mBERT and would need to generate English mapping from scratch. Did you learn the W matrix for English by aligning English to itself using MUSE? Wouldn't that be redundant?

Thanks,

Detailed setting info on anchor generation

This is not an issue, but a question. Could you tell me more about how your anchors were generated for your experiments? The paper says anchors were computed from the evaluation set (which amounts to 5% of the total CoNLL data), but I'd like to know more details - like exactly how many sentences were used, for each language, etc.

Thanks!

What layer is aligned and what the English alignment matrix for?

Hi,

I've read your article and the readme here and I'm a little confused about 2 things:

  1. You state

We provide the alignment of the first LSTM output of ELMo to English

and also:

Note that our alignments were done on the second layer of ELMo

My understanding is that "fist LSTM output" is not the second layer. Am I misunderstanding the quote? What layer do you align?

  1. If the alignment matrices aligned the models to the English model, what is the purpose of the English alignment matrix?

Thanks!

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.