• april
    link
    97 months ago

    It’s not exactly the same tech but it’s very similar. The Transformer architecture made a big difference. Before that we only had LSTMs which do sequence modelling in a different way that made far back things influence the result less.