SeeKer

tags
Transformers, GPT
paper
(Shuster et al. 2022)

Architecture

This is an extension that can be applied to any Transformer model by introducing “search”, “knowledge”, and “response” modules during pre-training of the model. It has the same applications as the base model it extends.

Parameter count

Depends on the base model being extended.

Bibliography

  1. . . March 29, 2022DOI.
Last changed | authored by

Comments

Loading comments...

Leave a comment

Back to Notes