--- library_name: setfit metrics: - accuracy pipeline_tag: text-classification tags: - setfit - sentence-transformers - text-classification - generated_from_setfit_trainer widget: [] inference: true license: apache-2.0 base_model: intfloat/multilingual-e5-large --- # TwinTransitionMapper_Green This repository contains the model for our paper entitled [Not all twins are identical: the digital layer of “twin” transition market applications](https://drive.google.com/file/d/1MN0GSl1FExHYkDyN_VhEt8yFwMX1MM4x/view?usp=drive_link) which is under review in Regional Studies (https://www.tandfonline.com/journals/cres20). This is a [SetFit](https://github.com/huggingface/setfit) model that can be used for Text Classification. This SetFit model uses [intfloat/multilingual-e5-large](https://huggingface.co/intfloat/multilingual-e5-large) as the Sentence Transformer embedding model. A [LogisticRegression](https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LogisticRegression.html) instance is used for classification. The model has been trained on paragraphs from German company websites using an efficient few-shot learning technique that involves: 1. Fine-tuning a [Sentence Transformer](https://www.sbert.net) with contrastive learning. 2. Training a classification head with features from the fine-tuned Sentence Transformer. The model is designed to predict the clean technology capabilities of German companies based on their website texts. It is intended to be used in conjunction with the [TwinTransitionMapper_AI](https://huggingface.co/LKriesch/TwinTransitionMapper_AI) model to identify companies contributing to the twin transition in Germany. For detailed information on the fine-tuning process and the results of these models, please refer to the [paper](https://drive.google.com/file/d/1MN0GSl1FExHYkDyN_VhEt8yFwMX1MM4x/view?usp=drive_link). ### Model Description - **Model Type:** SetFit - **Classification head:** a [LogisticRegression](https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LogisticRegression.html) instance - **Maximum Sequence Length:** 512 tokens ### Model Sources - **Repository:** [SetFit on GitHub](https://github.com/huggingface/setfit) - **Paper:** [Efficient Few-Shot Learning Without Prompts](https://arxiv.org/abs/2209.11055) - **Blogpost:** [SetFit: Efficient Few-Shot Learning Without Prompts](https://huggingface.co/blog/setfit) ## Uses ### Direct Use for Inference First install the SetFit library: ```bash pip install setfit ``` Then you can load this model and run inference. ```python from setfit import SetFitModel # Download from the 🤗 Hub model = SetFitModel.from_pretrained("LKriesch/TwinTransitionMapper_Green") # Run inference preds = model("I loved the spiderman movie!") ``` ## Training Details ### Framework Versions - Python: 3.9.19 - SetFit: 1.0.3 - Sentence Transformers: 3.0.1 - Transformers: 4.44.0 - PyTorch: 2.4.0+cu124 - Datasets: 2.16.1 - Tokenizers: 0.19.1 ## Citation ### BibTeX ```bibtex @article{https://doi.org/10.48550/arxiv.2209.11055, doi = {10.48550/ARXIV.2209.11055}, url = {https://arxiv.org/abs/2209.11055}, author = {Tunstall, Lewis and Reimers, Nils and Jo, Unso Eun Seo and Bates, Luke and Korat, Daniel and Wasserblat, Moshe and Pereg, Oren}, keywords = {Computation and Language (cs.CL), FOS: Computer and information sciences, FOS: Computer and information sciences}, title = {Efficient Few-Shot Learning Without Prompts}, publisher = {arXiv}, year = {2022}, copyright = {Creative Commons Attribution 4.0 International} } ```