[FOX-Ebook]Getting Started with Google BERT: Build and train state-of-the-art natural language ...

[FOX-Ebook]Getting Started with Google BERT: Build and train state-of-the-art natural language processing models using BERT
¥34.99 市场价 ¥899.99
库存
9999
数量
-
+
联系卖家   QQ:316821785   微信:zbook8_com  电话:13111111111   
商品特色:担保交易手动发货商品,工作人员手动发货。

自动发货宝贝:购买后直接到我买到的商品-订单详情-收货信息获取下载链接。
手动发货宝贝:购买后请留言邮箱或联系方式,0-4小时内由工作人员发到您邮箱。
购买后任何问题请联系商家或直接联系本站站务微信或者QQ。
书籍格式:
isbn:
排版:
新旧程度:

-------如果这里没有任何信息,不是真没有,是我们懒!请复制书名上amazon搜索书籍信息。-------

Kickstart your NLP journey by exploring BERT and its variants such as ALBERT, RoBERTa, DistilBERT, VideoBERT, and more with Hugging Face’s transformers library

Key Features

  • Explore the encoder and decoder of the transformer model
  • Become well-versed with BERT along with ALBERT, RoBERTa, and DistilBERT
  • Discover how to pre-train and fine-tune BERT models for several NLP tasks

Book Description

BERT (bidirectional encoder representations from transformer) has revolutionized the world of natural language processing (NLP) with promising results. This book is an introductory guide that will help you get to grips with Google’s BERT architecture. With a detailed explanation of the transformer architecture, this book will help you understand how the transformer’s encoder and decoder work.

You’ll explore the BERT architecture by learning how the BERT model is pre-trained and how to use pre-trained BERT for downstream tasks by fine-tuning it for NLP tasks such as sentiment analysis and text summarization with the Hugging Face transformers library. As you advance, you’ll learn about different variants of BERT such as ALBERT, RoBERTa, and ELECTRA, and look at SpanBERT, which is used for NLP tasks like question answering. You’ll also cover simpler and faster BERT variants based on knowledge distillation such as DistilBERT and TinyBERT. The book takes you through MBERT, XLM, and XLM-R in detail and then introduces you to sentence-BERT, which is used for obtaining sentence representation. Finally, you’ll discover domain-specific BERT models such as BioBERT and ClinicalBERT, and discover an interesting variant called VideoBERT.

By the end of this BERT book, you’ll be well-versed with using BERT and its variants for performing practical NLP tasks.

What you will learn

  • Understand the transformer model from the ground up
  • Find out how BERT works and pre-train it using masked language model (MLM) and next sentence prediction (NSP) tasks
  • Get hands-on with BERT by learning to generate contextual word and sentence embeddings
  • Fine-tune BERT for downstream tasks
  • Get to grips with ALBERT, RoBERTa, ELECTRA, and SpanBERT models
  • Get the hang of the BERT models based on knowledge distillation
  • Understand cross-lingual models such as XLM and XLM-R
  • Explore Sentence-BERT, VideoBERT, and BART

Who This Book Is For

This book is for NLP professionals and data scientists looking to simplify NLP tasks to enable efficient language understanding using BERT. A basic understanding of NLP concepts and deep learning is required to get the best out of this book.


暂无评价
暂时没有数据

交易规则

免责声明


1、本站所有分享材料(数据、资料)均为网友上传,如有侵犯您的任何权利,请您第一时间通过微信(zbook8_com) 、QQ(316821785)、 电话(13111111111)联系本站,本站将在24小时内回复您的诉求!谢谢!
2、本站所有商品,除特殊说明外,均为(电子版)Ebook,请购买分享内容前请务必注意。特殊商品有说明实物的,按照说明为准。

发货方式


1、自动:在上方保障服务中标有自动发货的宝贝,拍下后,将会自动收到来自卖家的宝贝获取(下载)链接   [个人中心->我的订单->点击订单 查看详情];
2、手动:未标有自动发货的的宝贝,拍下后,通过QQ或订单中的电话联系对方。

退款说明


1、描述:书籍描述(含标题)与实际不一致的(例:描述PDF,实际为epub、缺页少页、版本不符等);
2、链接:部分图书会给出链接,直接链接到官网或者其他站点,以便于提示,如与给出不符等;
3、发货:手动发货书籍,在卖家未发货前,已申请退款的;
4、其他:如质量方面的硬性常规问题等。
注:经核实符合上述任一,均支持退款,但卖家予以积极解决问题则除外。交易中的商品,卖家无法对描述进行修改!

注意事项


1、在未购买下前,双方在QQ上所商定的内容,亦可成为纠纷评判依据(商定与描述冲突时,商定为准);
2、在宝贝同时有网站演示与图片演示,且站演与图演不一致时,默认按图演作为纠纷评判依据(特别声明或有商定除外);
3、在没有"无任何正当退款依据"的前提下,写有"一旦售出,概不支持退款"等类似的声明,视为无效声明;
4、虽然交易产生纠纷的几率很小,但请尽量保留如聊天记录这样的重要信息,以防产生纠纷时便于网站工作人员介入快速处理。