<codeBook xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:xsd="http://www.w3.org/2001/XMLSchema" xsi:schemaLocation="ddi:codebook:2_5 http://www.ddialliance.org/Specification/DDI-Codebook/2.5/XMLSchema/codebook.xsd" xmlns="ddi:codebook:2_5">
  <docDscr>
    <citation>
      <titlStmt>
        <titl xml:lang="sv">Förtränade inbäddningar</titl>
        <parTitl xml:lang="en">Pretrained embeddings</parTitl>
        <IDNo agency="SND">doi-10-23695-ewz4-5030-0</IDNo>
        <IDNo agency="DOI">https://doi.org/10.23695/EWZ4-5030</IDNo>
      </titlStmt>
      <prodStmt>
        <producer xml:lang="en" abbr="SND">Swedish National Data Service</producer>
        <producer xml:lang="sv" abbr="SND">Svensk nationell datatjänst</producer>
      </prodStmt>
      <holdings URI="https://doi.org/10.23695/EWZ4-5030">Landing page</holdings>
    </citation>
  </docDscr>
  <stdyDscr>
    <citation>
      <titlStmt>
        <titl xml:lang="sv">Förtränade inbäddningar</titl>
        <parTitl xml:lang="en">Pretrained embeddings</parTitl>
        <IDNo agency="SND">doi-10-23695-ewz4-5030-0</IDNo>
        <IDNo agency="DOI">https://doi.org/10.23695/EWZ4-5030</IDNo>
      </titlStmt>
      <rspStmt>
        <AuthEnty xml:lang="en" affiliation="">Språkbanken Text</AuthEnty>
      </rspStmt>
      <prodStmt />
      <distStmt>
        <distrbtr xml:lang="en" abbr="SND" URI="https://snd.se">Swedish National Data Service</distrbtr>
        <distrbtr xml:lang="sv" abbr="SND" URI="https://snd.se">Svensk nationell datatjänst</distrbtr>
        <distDate xml:lang="en" date="2025-01-01" />
      </distStmt>
      <verStmt>
        <version elementVersion="0" elementVersionDate="2025-01-01" />
      </verStmt>
      <holdings URI="https://doi.org/10.23695/EWZ4-5030">Landing page</holdings>
    </citation>
    <stdyInfo>
      <subject />
      <abstract xml:lang="en" contentType="abstract">Embeddings (mappings of linguistic units, such as words, sentences, characters, to vectors of real numbers) are playing an extremely important role in modern language technology. Training the embedding models is often costly, which is why pretrained embeddings are widely used. On this page we provide lists of various pretrained embeddings for Swedish and of studies that focus on evaluating Swedish embeddings. If you have suggestions or comments, please contact us.
Embeddings

Facebook FastText models: Common Crawl + Wiki, Wiki, Wiki with cross-lingual alignment
NLPL repository: Word2Vec Continuous Skipgram (CoNLL17 corpus); ELMO (CoNLL17 corpus), ELMO (Wiki)
NLPLAB at Linköping University: a pretrained Word2Vec model (trained on a Göteborgs-Posten corpus); a script for training both cbow and sgns Word2Vec; a paper comparing Word2Vec and GloVe to Saldo
The National Library's (Kungliga bibliotekets) models: BERT, BERT fine-tuned for NER, ALBERT
The Public Employment Service's (Arbetsförmedlingens )models: BERT
Polyglot
Kyubyong Park's models: trained on Wiki, Word2Vec and FastText
Flair models. See also our Flair model for Swedish POS tagging.
Språkbanken Text diachronic embeddings.

Evaluation studies

Sahlgren, Magnus, and Fredrik Olsson. 2016. Gender Bias in Pretrained Swedish Embeddings. Proceedings of the 22nd Nordic Conference on Computational Linguistics.
Fallgren, Per, Jesper Segeblad, and Marco Kuhlmann. 2016. Towards a standard dataset of swedish word vectors. Sixth Swedish Language Technology Conference (SLTC).
Holmer, Daniel. 2020. Context matters: Classifying Swedish texts using BERT's deep bidirectional word embeddings. Bachelor thesis at Linköping University.
Adewumi, Tosin, Foteini Liwicki and Marcus Liwicki. 2020. Exploring Swedish &amp; English fastText Embeddings with the Transformer
Adewumi, Tosin, Foteini Liwicki and Marcus Liwicki. 2020. Corpora Compared: The Case of the Swedish Gigaword &amp; Wikipedia Corpora</abstract>
      <abstract xml:lang="sv" contentType="abstract">Embeddings (mappings of linguistic units, such as words, sentences, characters, to vectors of real numbers) are playing an extremely important role in modern language technology. Training the embedding models is often costly, which is why pretrained embeddings are widely used. On this page we provide lists of various pretrained embeddings for Swedish and of studies that focus on evaluating Swedish embeddings. If you have suggestions or comments, please contact us.
Embeddings

Facebook FastText models: Common Crawl + Wiki, Wiki, Wiki with cross-lingual alignment
NLPL repository: Word2Vec Continuous Skipgram (CoNLL17 corpus); ELMO (CoNLL17 corpus), ELMO (Wiki)
NLPLAB at Linköping University: a pretrained Word2Vec model (trained on a Göteborgs-Posten corpus); a script for training both cbow and sgns Word2Vec; a paper comparing Word2Vec and GloVe to Saldo
The National Library's (Kungliga bibliotekets) models: BERT, BERT fine-tuned for NER, ALBERT
The Public Employment Service's (Arbetsförmedlingens )models: BERT
Polyglot
Kyubyong Park's models: trained on Wiki, Word2Vec and FastText
Flair models. See also our Flair model for Swedish POS tagging.
Språkbanken Text diachronic embeddings.

Evaluation studies

Sahlgren, Magnus, and Fredrik Olsson. 2016. Gender Bias in Pretrained Swedish Embeddings. Proceedings of the 22nd Nordic Conference on Computational Linguistics.
Fallgren, Per, Jesper Segeblad, and Marco Kuhlmann. 2016. Towards a standard dataset of swedish word vectors. Sixth Swedish Language Technology Conference (SLTC).
Holmer, Daniel. 2020. Context matters: Classifying Swedish texts using BERT's deep bidirectional word embeddings. Bachelor thesis at Linköping University.
Adewumi, Tosin, Foteini Liwicki and Marcus Liwicki. 2020. Exploring Swedish &amp; English fastText Embeddings with the Transformer
Adewumi, Tosin, Foteini Liwicki and Marcus Liwicki. 2020. Corpora Compared: The Case of the Swedish Gigaword &amp; Wikipedia Corpora</abstract>
      <sumDscr />
    </stdyInfo>
    <method>
      <dataColl />
    </method>
    <dataAccs>
      <useStmt>
        <restrctn xml:lang="en">Access to data through an external actor. </restrctn>
        <restrctn xml:lang="sv">Åtkomst till data via extern aktör. </restrctn>
      </useStmt>
    </dataAccs>
    <othrStdyMat />
  </stdyDscr>
</codeBook>