Hands-on question answering systems with BERT : (Record no. 91153)

MARC details
000 -LEADER
fixed length control field 02006nam a22001817a 4500
008 - FIXED-LENGTH DATA ELEMENTS--GENERAL INFORMATION
fixed length control field 231122b2022 |||||||| |||| 00| 0 eng d
020 ## - INTERNATIONAL STANDARD BOOK NUMBER
International Standard Book Number 9781484283912
082 ## - DEWEY DECIMAL CLASSIFICATION NUMBER
Classification number 006.32 SAB-N
100 ## - MAIN ENTRY--PERSONAL NAME
Personal name Sabharwal, Navin
245 ## - TITLE STATEMENT
Title Hands-on question answering systems with BERT :
Remainder of title applications n neural networks and natural language processing /
Statement of responsibility, etc. Navin Sabharwal and Amit Agrawal
260 ## - PUBLICATION, DISTRIBUTION, ETC. (IMPRINT)
Place of publication, distribution, etc. New York
Name of publisher, distributor, etc. Springer
Date of publication, distribution, etc. 2022
300 ## - PHYSICAL DESCRIPTION
Extent 184p.
500 ## - GENERAL NOTE
General note Get hands-on knowledge of how BERT (Bidirectional Encoder Representations from Transformers) can be used to develop question answering (QA) systems by using natural language processing (NLP) and deep learning. The book begins with an overview of the technology landscape behind BERT. It takes you through the basics of NLP, including natural language understanding with tokenization, stemming, and lemmatization, and bag of words. Next, you'll look at neural networks for NLP starting with its variants such as recurrent neural networks, encoders and decoders, bi-directional encoders and decoders, and transformer models. Along the way, you'll cover word embedding and their types along with the basics of BERT. After this solid foundation, you'll be ready to take a deep dive into BERT algorithms such as masked language models and next sentence prediction. You'll see different BERT variations followed by a hands-on example of a question answering system. Hands-on Question Answering Systems with BERT is a good starting point for developers and data scientists who want to develop and design NLP systems using BERT. It provides step-by-step guidance for using BERT. You will: Examine the fundamentals of word embeddings Apply neural networks and BERT for various NLP tasks Develop a question-answering system from scratch Train question-answering systems for your own data
650 ## - SUBJECT ADDED ENTRY--TOPICAL TERM
Topical term or geographic name as entry element Machine learning
650 ## - SUBJECT ADDED ENTRY--TOPICAL TERM
Topical term or geographic name as entry element Computer software
650 ## - SUBJECT ADDED ENTRY--TOPICAL TERM
Topical term or geographic name as entry element Natural language processing (Computer science)
650 ## - SUBJECT ADDED ENTRY--TOPICAL TERM
Topical term or geographic name as entry element Neural networks (Computer science)
952 ## - LOCATION AND ITEM INFORMATION (KOHA)
Withdrawn status
Holdings
Lost status Source of classification or shelving scheme Damaged status Not for loan Collection code Current library Shelving location Date acquired Total Checkouts Full call number Barcode Date last seen Koha item type
  Dewey Decimal Classification     003-007 BITS Pilani Hyderabad General Stack (For lending) 22/11/2023   006.32 SAB-N 47516 13/07/2024 Books
An institution deemed to be a University Estd. Vide Sec.3 of the UGC
Act,1956 under notification # F.12-23/63.U-2 of Jun 18,1964

© 2024 BITS-Library, BITS-Hyderabad, India.