Jump to navigation
Who We Are
Policies & Guidelines
Learn at LTI
Explore Our Work
Back to the catalogue
If you have a problem in filling in this form, contact lti.catalogue AT gmail.com.
Fields marked with (
) are required.
The email is only for internal purposes.
the submission will not be considered until you confirm that you own this email.
You will receive a confirmation email upon submitting. The confirmation email might get into your junk email, so check your SPAM folder. If you do not receive an email, please contact us.
A proof for not being spam
This information will appear in the catalogue.
Natural Language Processing/Computational Linguistics
Information Retrieval, Text Mining and Analytics
Spoken Interfaces and Dialogue Processing
Keywords (comma separated, internal and not shown to public)
Direct Download Link
(If you provide a direct download link, please also provide the IP agreement, and the Required Acknowledgement)
<p>Nathan Schneider, Noah A. Smith, et al.</p>
<p><strong>AMALGrAM</strong> (<strong>A</strong> <strong>M</strong>achine <strong>A</strong>nalyzer of <strong>L</strong>exical <strong>Gr</strong>oupings <strong>A</strong>nd <strong>M</strong>eanings) analyzes English sentences for multiword expressions (MWEs) and noun and verb supersenses. For example, given the sentence</p> <blockquote> <p>I do n't think he 's afraid to take a strong stand on gun control , what with his upbringing in El Paso .</p> </blockquote> <p>the analysis</p> <pre><code>I do|`a n't think|cognition he 's|stative afraid to take_a_ strong _stand|cognition on gun_control|ARTIFACT , what_with his upbringing|ATTRIBUTE in El_Paso|LOCATION . </code></pre> <p>will be predicted, grouping "take a stand", "gun control", "what with", and "El Paso" as MWEs and labeling several lexical expressions with supersenses (UPPERCASE for nouns, lowercase for verbs). The model and algorithms implemented in the tool are described in Schneider et al. (<em>TACL</em> 2014, NAACL-HLT 2015), and resources required to use it are available at <a href="http://www.ark.cs.cmu.edu/LexSem/">http://www.ark.cs.cmu.edu/LexSem/</a>.</p> <p>More generally, the codebase supports supervised discriminative learning and structured prediction of statistical sequence models over discrete data, i.e., taggers. It implements the structured perceptron (Collins, EMNLP 2002) for learning and the Viterbi algorithm for decoding. Cython is used to make decoding reasonably fast, even with millions of features.</p>
Availability (e.g. source code, binary only, XML file, etc.)
Support Status (e.g. as-is, maintained, etc.)
Prerequisites (e.g. Windows XP, Java 1.6, etc.)
<p>Python 2.7 with Cython</p> <p>Has been tested on Unix platforms</p>
Required Acknowledgement (e.g. paper to cite)
<p>The original MWE identification system is described in:</p> <ul> <li><strong><a href="http://www.cs.cmu.edu/%7Eark/LexSem/mwe.pdf">Discriminative Lexical Semantic Segmentation with Gaps: Running the MWE Gamut</a></strong> <strong><a style="color: #ff0080;" href="http://www.cs.cmu.edu/%7Enschneid/#pubs_tacl2014">[errata]</a></strong><br /> <a href="http://nathan.cl">Nathan Schneider</a>, <a href="https://github.com/ecdanchik">Emily Danchik</a>, <a href="http://www.cs.cmu.edu/%7Ecdyer/">Chris Dyer</a>, and <a href="http://www.cs.cmu.edu/%7Enasmith/">Noah A. Smith</a>.<br /> In <em>Transactions of the Association for Computational Linguistics</em>, 2(April):193−206, 2014.</li> </ul> <p>The supersense annotations and the combined MWE+supersense tagger are described in:</p> <ul> <li><strong> <a href="http://www.cs.cmu.edu/%7Eark/LexSem/sst.pdf">A Corpus and Model Integrating Multiword Expressions and Supersenses</a></strong><br /> <a href="http://nathan.cl">Nathan Schneider</a> and <a href="http://www.cs.cmu.edu/%7Enasmith/">Noah A. Smith</a>.<br /> In <em>Proceedings of the 2015 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</em>, Denver, Colorado, June 2015.</li> </ul>
might be helpful)
<pre>The software is released under the GPLv3 license.</pre>
Contact (e.g. e-mail)
Additional Comments (internal and not shown to public)