|
|
|
|
LEADER |
03878cam a2200613 4500 |
001 |
183-1793559562 |
003 |
DE-627 |
005 |
20240319082454.0 |
007 |
tu |
008 |
220222s2022 xxu||||| 00| ||eng c |
010 |
|
|
|a 2021060635
|
020 |
|
|
|a 9781316519332
|c hardback
|9 978-1-316-51933-2
|
024 |
3 |
|
|a 9781316519332
|
024 |
7 |
|
|a 10.1017/9781009023405
|2 doi
|
035 |
|
|
|a (DE-627)1793559562
|
035 |
|
|
|a (DE-599)KXP1793559562
|
035 |
|
|
|a (OCoLC)1319077261
|
040 |
|
|
|a DE-627
|b ger
|c DE-627
|e rda
|
041 |
|
|
|a eng
|
044 |
|
|
|c XD-US
|
050 |
|
0 |
|a Q325.73
|
084 |
|
|
|a 68-01
|2 msc
|
084 |
|
|
|a 68T07
|2 msc
|
084 |
|
|
|a 54.72
|2 bkl
|
100 |
1 |
|
|a Roberts, Daniel A.
|d 1987-
|e VerfasserIn
|0 (DE-588)1261776224
|0 (DE-627)1809078989
|4 aut
|
245 |
1 |
4 |
|a The principles of deep learning theory
|b an effective theory approach to understanding neural networks
|c Daniel A. Roberts (MIT), Sho Yaida (Meta AI) ; based on research in collaboration with Boris Hanin (Princeton University)
|
264 |
|
1 |
|a Cambridge
|a New York, NY
|a Port Melbourne, VIC
|a New Delhi
|a Singapore
|b Cambridge University Press
|c 2022
|
300 |
|
|
|a x, 460 Seiten
|b Diagramme
|
336 |
|
|
|a Text
|b txt
|2 rdacontent
|
337 |
|
|
|a ohne Hilfsmittel zu benutzen
|b n
|2 rdamedia
|
338 |
|
|
|a Band
|b nc
|2 rdacarrier
|
500 |
|
|
|a Includes bibliographical references and index
|
520 |
|
|
|a "This textbook establishes a theoretical framework for understanding deep learning models of practical relevance. With an approach that borrows from theoretical physics, Roberts and Yaida provide clear and pedagogical explanations of how realistic deep neural networks actually work. To make results from the theoretical forefront accessible, the authors eschew the subject's traditional emphasis on intimidating formality without sacrificing accuracy. Straightforward and approachable, this volume balances detailed first-principle derivations of novel results with insight and intuition for theorists and practitioners alike. This self-contained textbook is ideal for students and researchers interested in artificial intelligence with minimal prerequisites of linear algebra, calculus, and informal probability theory, and it can easily fill a semester-long course on deep learning theory. For the first time, the exciting practical advances in modern artificial intelligence capabilities can be matched with a set of effective principles, providing a timeless blueprint for theoretical research in deep learning"--
|
650 |
|
0 |
|a Deep learning (Machine learning)
|
650 |
|
4 |
|a SCIENCE / Physics / Mathematical & Computational
|
689 |
0 |
0 |
|D s
|0 (DE-588)1135597375
|0 (DE-627)890512922
|0 (DE-576)489847412
|a Deep learning
|2 gnd
|
689 |
0 |
|
|5 (DE-627)
|
700 |
1 |
|
|a Yaida, Sho
|e VerfasserIn
|0 (DE-588)1261764668
|0 (DE-627)1809045266
|4 aut
|
700 |
1 |
|
|a Hanin, Boris
|e MitwirkendeR
|0 (DE-588)1261776534
|0 (DE-627)1809079454
|4 ctb
|
776 |
1 |
|
|z 9781009023405
|c epub
|
776 |
0 |
8 |
|i Erscheint auch als
|n Online-Ausgabe
|a Roberts, Daniel A., 1987 -
|t The principles of deep learning theory
|d Cambridge : Cambridge University Press, 2022
|h 1 online resource (x, 460 pages)
|w (DE-627)1807444929
|z 9781009023405
|
856 |
4 |
2 |
|u https://zbmath.org/1507.68003
|y zbMATH
|z Rezension
|
924 |
0 |
|
|a 4136517947
|b DE-Hil2
|9 Hil 2
|c GBV
|d c
|g CSC 726 : R54
|
924 |
0 |
|
|a 4156789071
|b DE-Hil2
|9 Hil 2
|c GBV
|d d
|g BWL 05
|h HA
|
924 |
0 |
|
|a 4437782011
|b DE-700
|9 700
|c GBV
|d c
|g BFL M / Rob
|h N
|
924 |
0 |
|
|a 4501538066
|b DE-517
|9 517
|c GBV
|d c
|h 1608
|
924 |
0 |
|
|a 4205343159
|b DE-B1532
|9 B 1532
|c GBV
|d d
|g 2022/620
|
924 |
0 |
|
|a 4152851600
|b DE-25
|9 25
|c BSZ
|d c
|g FX 2022/257
|
924 |
0 |
|
|a 422165435X
|b DE-291-406
|9 291/406
|c BSZ
|d b
|
924 |
0 |
|
|a 4413923510
|b DE-Zi4
|9 Zi 4
|c BSZ
|d c
|
924 |
0 |
|
|a 4252952506
|b DE-Frei3c
|9 Frei 3c
|c BSZ
|d b
|g M/Roberts, D
|
936 |
b |
k |
|a 54.72
|j Künstliche Intelligenz
|0 (DE-627)10641240X
|
951 |
|
|
|a BO
|
980 |
|
|
|a 1793559562
|b 183
|c sid-183-col-kxpbbi
|
SOLR
_version_ |
1797788619810799616 |
author |
Roberts, Daniel A., Yaida, Sho |
author2 |
Hanin, Boris |
author2_role |
ctb |
author2_variant |
b h bh |
author_facet |
Roberts, Daniel A., Yaida, Sho, Hanin, Boris |
author_role |
aut, aut |
author_sort |
Roberts, Daniel A. 1987- |
author_variant |
d a r da dar, s y sy |
building |
Library A |
callnumber-first |
Q - Science |
callnumber-label |
Q325 |
callnumber-raw |
Q325.73 |
callnumber-search |
Q325.73 |
callnumber-sort |
Q 3325.73 |
callnumber-subject |
Q - General Science |
collection |
sid-183-col-kxpbbi |
contents |
"This textbook establishes a theoretical framework for understanding deep learning models of practical relevance. With an approach that borrows from theoretical physics, Roberts and Yaida provide clear and pedagogical explanations of how realistic deep neural networks actually work. To make results from the theoretical forefront accessible, the authors eschew the subject's traditional emphasis on intimidating formality without sacrificing accuracy. Straightforward and approachable, this volume balances detailed first-principle derivations of novel results with insight and intuition for theorists and practitioners alike. This self-contained textbook is ideal for students and researchers interested in artificial intelligence with minimal prerequisites of linear algebra, calculus, and informal probability theory, and it can easily fill a semester-long course on deep learning theory. For the first time, the exciting practical advances in modern artificial intelligence capabilities can be matched with a set of effective principles, providing a timeless blueprint for theoretical research in deep learning"-- |
ctrlnum |
(DE-627)1793559562, (DE-599)KXP1793559562, (OCoLC)1319077261 |
doi_str_mv |
10.1017/9781009023405 |
facet_avail |
Local |
facet_local_del330 |
Deep learning |
finc_class_facet |
Allgemeine Naturwissenschaft |
fincclass_txtF_mv |
science-computerscience |
footnote |
Includes bibliographical references and index |
format |
Book |
format_access_txtF_mv |
Book, E-Book |
format_de14 |
Book, E-Book |
format_de15 |
Book, E-Book |
format_del152 |
Buch |
format_detail_txtF_mv |
text-print-monograph-independent |
format_dezi4 |
e-Book |
format_finc |
Book, E-Book |
format_legacy |
Book |
format_legacy_nrw |
Book, E-Book |
format_nrw |
Book, E-Book |
format_strict_txtF_mv |
Book |
geogr_code |
not assigned |
geogr_code_person |
not assigned |
id |
183-1793559562 |
illustrated |
Not Illustrated |
imprint |
Cambridge, New York, NY, Port Melbourne, VIC, New Delhi, Singapore, Cambridge University Press, 2022 |
imprint_str_mv |
Cambridge; New York, NY; Port Melbourne, VIC; New Delhi; Singapore: Cambridge University Press, 2022 |
institution |
FID-BBI-DE-23 |
is_hierarchy_id |
|
is_hierarchy_title |
|
isbn |
9781316519332 |
isbn_isn_mv |
9781009023405 |
language |
English |
last_indexed |
2024-04-30T19:20:32.594Z |
lccn |
2021060635 |
marc024a_ct_mv |
9781316519332, 10.1017/9781009023405 |
marc_error |
[geogr_code]Unable to make public java.lang.AbstractStringBuilder java.lang.AbstractStringBuilder.append(java.lang.String) accessible: module java.base does not "opens java.lang" to unnamed module @64e01542 |
match_str |
roberts2022theprinciplesofdeeplearningtheoryaneffectivetheoryapproachtounderstandingneuralnetworks |
mega_collection |
K10plus Verbundkatalog |
oclc_num |
1319077261 |
physical |
x, 460 Seiten; Diagramme |
publishDate |
2022 |
publishDateSort |
2022 |
publishPlace |
Cambridge |
publisher |
Cambridge University Press |
record_format |
marcfinc |
record_id |
1793559562 |
recordtype |
marcfinc |
rvk_facet |
No subject assigned |
source_id |
183 |
spelling |
Roberts, Daniel A. 1987- VerfasserIn (DE-588)1261776224 (DE-627)1809078989 aut, The principles of deep learning theory an effective theory approach to understanding neural networks Daniel A. Roberts (MIT), Sho Yaida (Meta AI) ; based on research in collaboration with Boris Hanin (Princeton University), Cambridge New York, NY Port Melbourne, VIC New Delhi Singapore Cambridge University Press 2022, x, 460 Seiten Diagramme, Text txt rdacontent, ohne Hilfsmittel zu benutzen n rdamedia, Band nc rdacarrier, Includes bibliographical references and index, "This textbook establishes a theoretical framework for understanding deep learning models of practical relevance. With an approach that borrows from theoretical physics, Roberts and Yaida provide clear and pedagogical explanations of how realistic deep neural networks actually work. To make results from the theoretical forefront accessible, the authors eschew the subject's traditional emphasis on intimidating formality without sacrificing accuracy. Straightforward and approachable, this volume balances detailed first-principle derivations of novel results with insight and intuition for theorists and practitioners alike. This self-contained textbook is ideal for students and researchers interested in artificial intelligence with minimal prerequisites of linear algebra, calculus, and informal probability theory, and it can easily fill a semester-long course on deep learning theory. For the first time, the exciting practical advances in modern artificial intelligence capabilities can be matched with a set of effective principles, providing a timeless blueprint for theoretical research in deep learning"--, Deep learning (Machine learning), SCIENCE / Physics / Mathematical & Computational, s (DE-588)1135597375 (DE-627)890512922 (DE-576)489847412 Deep learning gnd, (DE-627), Yaida, Sho VerfasserIn (DE-588)1261764668 (DE-627)1809045266 aut, Hanin, Boris MitwirkendeR (DE-588)1261776534 (DE-627)1809079454 ctb, 9781009023405 epub, Erscheint auch als Online-Ausgabe Roberts, Daniel A., 1987 - The principles of deep learning theory Cambridge : Cambridge University Press, 2022 1 online resource (x, 460 pages) (DE-627)1807444929 9781009023405, https://zbmath.org/1507.68003 zbMATH Rezension |
spellingShingle |
Roberts, Daniel A., Yaida, Sho, The principles of deep learning theory: an effective theory approach to understanding neural networks, "This textbook establishes a theoretical framework for understanding deep learning models of practical relevance. With an approach that borrows from theoretical physics, Roberts and Yaida provide clear and pedagogical explanations of how realistic deep neural networks actually work. To make results from the theoretical forefront accessible, the authors eschew the subject's traditional emphasis on intimidating formality without sacrificing accuracy. Straightforward and approachable, this volume balances detailed first-principle derivations of novel results with insight and intuition for theorists and practitioners alike. This self-contained textbook is ideal for students and researchers interested in artificial intelligence with minimal prerequisites of linear algebra, calculus, and informal probability theory, and it can easily fill a semester-long course on deep learning theory. For the first time, the exciting practical advances in modern artificial intelligence capabilities can be matched with a set of effective principles, providing a timeless blueprint for theoretical research in deep learning"--, Deep learning (Machine learning), SCIENCE / Physics / Mathematical & Computational, Deep learning |
title |
The principles of deep learning theory: an effective theory approach to understanding neural networks |
title_auth |
The principles of deep learning theory an effective theory approach to understanding neural networks |
title_full |
The principles of deep learning theory an effective theory approach to understanding neural networks Daniel A. Roberts (MIT), Sho Yaida (Meta AI) ; based on research in collaboration with Boris Hanin (Princeton University) |
title_fullStr |
The principles of deep learning theory an effective theory approach to understanding neural networks Daniel A. Roberts (MIT), Sho Yaida (Meta AI) ; based on research in collaboration with Boris Hanin (Princeton University) |
title_full_unstemmed |
The principles of deep learning theory an effective theory approach to understanding neural networks Daniel A. Roberts (MIT), Sho Yaida (Meta AI) ; based on research in collaboration with Boris Hanin (Princeton University) |
title_short |
The principles of deep learning theory |
title_sort |
the principles of deep learning theory an effective theory approach to understanding neural networks |
title_sub |
an effective theory approach to understanding neural networks |
title_unstemmed |
The principles of deep learning theory: an effective theory approach to understanding neural networks |
topic |
Deep learning (Machine learning), SCIENCE / Physics / Mathematical & Computational, Deep learning |
topic_facet |
Deep learning (Machine learning), SCIENCE / Physics / Mathematical & Computational, Deep learning |
url |
https://zbmath.org/1507.68003 |
work_keys_str_mv |
AT robertsdaniela theprinciplesofdeeplearningtheoryaneffectivetheoryapproachtounderstandingneuralnetworks, AT yaidasho theprinciplesofdeeplearningtheoryaneffectivetheoryapproachtounderstandingneuralnetworks, AT haninboris theprinciplesofdeeplearningtheoryaneffectivetheoryapproachtounderstandingneuralnetworks, AT robertsdaniela principlesofdeeplearningtheoryaneffectivetheoryapproachtounderstandingneuralnetworks, AT yaidasho principlesofdeeplearningtheoryaneffectivetheoryapproachtounderstandingneuralnetworks, AT haninboris principlesofdeeplearningtheoryaneffectivetheoryapproachtounderstandingneuralnetworks |