oserikov / bigscience-gsoc-ideas Goto Github PK
View Code? Open in Web Editor NEWHere we track the Ideas for the GSoC.
Here we track the Ideas for the GSoC.
difficulty: scalable, can be both 170 and 340 hours
mentor: @oserikov , TBD
requirements:
useful links:
WIP
WIP
difficulty: scalable, can be both 170 and 340 hours
mentor: @oserikov , TBD
requirements:
useful links:
While HuggingFace quickly became the standard way to publish language models, several architectural trade-offs have been made to support the quick growth of the models' zoo. This resulted in several theoretically similar models being implemented by different teams, thus e.g. several alternative implementations of self-attentive transformers arose. While refactoring the whole zoo of models seems to be far from the accessible task, the interpretability community is forced to provide unification wrappers for handling such dissimilarities in similar models. The task is to provide a reasonable trade-off with the refactoring of the crucial models and providing the unified wrappers, and thus bring the unified interpretability API to the crucial HuggingFace models.
We could see this task from two prospects. First, one could unify the interpretability API of the sibling models such as BERT and RoBERTa . Second, one could think about bringing the unified interface to interpret and compare encoder models with e.g. encoder-decoder ones, allowing to study the similarities and distinctiveness in their behavior.
WIP
difficulty: scalable, can be both 170 and 340 hours
mentor: @oserikov , TBD
requirements:
useful links:
While HuggingFace quickly became the standard way to publish language models, several architectural trade-offs have been made to support the quick growth of the models' zoo. This resulted in several theoretically similar models being implemented by different teams, thus e.g. several alternative implementations of self-attentive transformers arose. While refactoring the whole zoo of models seems to be far from the accessible task, the interpretability community is forced to provide unification wrappers for handling such dissimilarities in similar models. The task is to provide a reasonable trade-off with the refactoring of the crucial models and providing the unified wrappers, and thus bring the unified interpretability API to the crucial HuggingFace models.
We could see this task from two prospects. First, one could unify the interpretability API of the sibling models such as BERT and RoBERTa . Second, one could think about bringing the unified interface to interpret and compare encoder models with e.g. encoder-decoder ones, allowing to study the similarities and distinctiveness in their behavior.
WIP
difficulty: scalable, can be both 170 and 340 hours
mentor: @oserikov , TBD
requirements:
useful links:
In Circuits, several abstract structures found in CV models were summarized. The Branches Specialization tendency of the CV neural networks as well as the Weight Banding property of NNs last layers have not been directly studied in LLms, though the findings of several papers (1, 2) could be related.
The task is to perform a study of the abstract structures representedness in CV and NLP models, by applying the same inspection techniques to both groups of models.
Reproduce the Branches Specialization test on some CV model; Reproduce the Individual Neurons analysis on BERT model.
difficulty: scalable, can be both 170 and 340 hours
mentor: @oserikov , TBD
requirements:
useful links:
WIP
WIP
difficulty: scalable, can be both 170 and 340 hours
mentor: @oserikov , TBD
requirements:
useful links:
During the season 2021/22, the BigScience team reached several crucial milestones by producing large-scale transformer language models. Some of them even come with the training checkpoints archived, thus allowing to study the emergence of the structures in language models. During this task, we propose to cover the released models with the supplementary interpretability information by applying classical XAI and probing methods described in the attached papers.
To better feel what the interpretability work looks like, we ask you to perform a diagnostic classification study of the GPT-like language model, using the SentEval data. Reach out to mentors as soon as possible to discuss the analysis results.
A declarative, efficient, and flexible JavaScript library for building user interfaces.
๐ Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. ๐๐๐
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google โค๏ธ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.