moonisali Goto Github PK
Name: Moonis Ali
Type: User
Bio: trying to learn, work in data science led disciplines. Mostly in social data science +linguistics/language domains like NLP/NLU
Name: Moonis Ali
Type: User
Bio: trying to learn, work in data science led disciplines. Mostly in social data science +linguistics/language domains like NLP/NLU
ML Zoomcamp fall 2021 homework and stuff
models and evaluation framework for trending topics detection
Plug and play modules to optimize the performances of your AI systems 🚀
Neural machine translation (NMT) is an approach to machine translation that uses an artificial neural network to predict the likelihood of a sequence of words, typically modeling entire sentences in a single integrated model. NMT departs from phrase-based statistical approaches that use separately engineered subcomponents.Neural machine translation (NMT) is not a drastic step beyond what has been traditionally done in statistical machine translation (SMT). Its main departure is the use of vector representations ("embeddings", "continuous space representations") for words and internal states. The structure of the models is simpler than phrase-based models. There is no separate language model, translation model, and reordering model, but just a single sequence model that predicts one word at a time. However, this sequence prediction is conditioned on the entire source sentence and the entire already produced target sequence. NMT models use deep learning and representation learning. The word sequence modeling was at first typically done using a recurrent neural network (RNN). A bidirectional recurrent neural network, known as an encoder, is used by the neural network to encode a source sentence for a second RNN, known as a decoder, that is used to predict words in the target language.Recurrent neural networks face difficulties in encoding long inputs into a single vector. This can be compensated by an attention mechanism which allows the decoder to focus on different parts of the input while generating each word of the output. There are further Coverage Models addressing the issues in such attention mechanisms, such as ignoring of past alignment information leading to over-translation and under-translation. Convolutional Neural Networks (Convnets) are in principle somewhat better for long continuous sequences, but were initially not used due to several weaknesses. These were successfully compensated for in 2017 by using "attention mechanisms". A attention-based model, the transformer architecture remains the dominant architecture for several language pairs.
Project developed during the SICSS-Aachen-Graz (July 2022) investigating the variation of sentiment in news articles tweeted by @BBCBreaking
Using transformers to predict next word and predict <mask> word
Saarland University NNTI WS2021 NLP Final Project
Notebooks from YouTube videos
includes jupyter notebooks used for data analysis in the project and blog post
OCTIS: Comparing Topic Models is Simple! A python package to optimize and evaluate topic models (accepted at EACL2021 demo track)
Omdena London Chapter's project : NLP to prevent self-harm for London teens
Examples and guides for using the OpenAI API
Discover Social Circles in Twitter Ego Network
ML Observability in a Notebook - Uncover Insights, Surface Problems, Monitor, and Fine Tune your Generative LLM, CV and Tabular Models
PiML (Python Interpretable Machine Learning) toolbox for model development and model validation
Code for the Political Volatility paper (Chico and Scott)
This repository contains the material of the practical section of course: "Introduction to Natural Language Processing" as a part of BICS(Bachelor in Computer Science) program at university of Luxembourg. These notebooks where created by me and my colleage https://github.com/katyamatya/ .
This package features data-science related tasks for developing new recognizers for Presidio. It is used for the evaluation of the entire system, as well as for evaluating specific PII recognizers or PII detection models.
Repository for Programming Assignment 2 for R Programming on Coursera
Open-source tools for prompt testing and experimentation
Python implementation of TextRank algorithms ("textgraphs") for phrase extraction
Code and data accompanying Natural Language Processing with PyTorch published by O'Reilly Media https://nlproc.info
Large scale font independent printed Urdu text data set
Blazing fast framework for fine-tuning similarity learning models
This repository contains a few simple projects with forms.
A declarative, efficient, and flexible JavaScript library for building user interfaces.
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google ❤️ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.