text
stringlengths 1
1k
⌀ | title
stringclasses 230
values |
---|---|
to fully enable SSL’s potential, and (iii) the absence of a unified vocabulary and theoretical
view of SSL. As SSL established a distinct paradigm from traditional reconstruction-based
unsupervised learning methods such as (denoising, variational) Autoencoders [Vincent
et al., 2008, 2010, Kingma and Welling, 2013], our vocabulary for understanding SSL
in a unified view is limited. In fact, attempts at unifying SSL methods under a single
viewpoint have only started to emerge in the last year [HaoChen et al., 2021, Balestriero
and LeCun, 2022, Shwartz-Ziv et al., 2022, Garrido et al., 2022b]. Without a common
ground to characterize the different components of SSL methods, it’s more challenging for
researchers to start working on SSL methods. Meanwhile, SSL research is in dire need for
new researchers since SSL is now deployed throughout the real-world. Yet, many open
research questions remain regarding SSL’s generalization guarantees, fairness proper- | A Cookbook of Self-Supervised Learning |
""" Write a function to find the largest integers from a given list of
numbers using heap queue algorithm. """
import heapq as hq
def heap_queue_largest(nums,n):
largest_nums = hq.nlargest(n, nums)
return largest_nums
### Task End ###
58
### Task Start ###
# These are the assertions for your function:
<insert assertions and problem description here>
C.2 Simple Feedback Prompt (6-shot)
# Write Python function to complete the task and pass the assertion tests.
### Task Start ###
# These are the assertions for your function:
assert count_ways(2) == 3
""" Write a function to find the number of ways to fill it with 2 x 1
dominoes for the given 3 x n board. """
def count_ways(n):
if n == 0:
return 1
if n == 1:
return 1
if n == 2:
return 3
return count_ways(n-1) + count_ways(n-2)
Feedback: The code above is wrong. Please fix it.
def count_ways(n):
A = [0] * (n + 1)
B = [0] * (n + 1)
A[0] = 1
A[1] = 0
B[0] = 0
B[1] = 1
for i in range(2, n+1): | Teaching Large Language Models to Self-Debug |
Pythia: A Suite for Analyzing Large Language Models
hurt performance at smaller scales, we find that our models
perform the same as equi-parameter OPT models across all
scales. We discuss areas where our results contradict widely
accepted maxims for training LLMs in Section 2.6.
2.1. Requirements for a Scientific Suite of LLMs
Pythia is envisioned as a suite for enabling and empowering
scientific research on the capacities and limitations of large
language models. After surveying the existing literature, we
found no existing suites of models which satisfied all the
following conditions:
Public Access Models are publicly released and are
trained on publicly available data.
Training Provenance
Intermediate checkpoints are avail-
able for analysis, all models are trained with the same data
ordering, and intermediate checkpoints can be linked with
the exact data seen up to that checkpoint. Training pro-
cedure as well as model and training hyperparameters are
well-documented. | Pythia- A Suite for Analyzing Large Language Models Across Training and Scaling |
Library of Congress Cataloging-in-Publication Data
names: Persily, Nathaniel, editor. | Tucker, Joshua A. (Joshua Aaron), 1971– editor.
title: Social media and democracy : the state of the field, prospects for reform / edited by Nathaniel
Persily, Joshua A. Tucker.
description: Cambridge, United Kingdom ; New York, NY : Cambridge University Press, 2020. | Series:
SSRC anxieties of democracy | Includes bibliographical references and index.
identifiers: lccn 2020013248 (print) | lccn 2020013249 (ebook) | isbn 9781108835558 (hardback) |
isbn 9781108890960 (ebook)
subjects: lcsh: Social media – Political aspects. | Online social networks – Political aspects. | Information
society – Political aspects. | Information technology – Political aspects. | Democracy. | Political
participation – Technological innovations.
classification: lcc hm742 .s628164 2020 (print) | lcc hm742 (ebook) |
ddc 302.23/1–dc23
LC record available at https://lccn.loc.gov/2020013248 | Social_Media_and_Democracy |
Hallucination in Conversation. EMNLP (2021).
[169] Haoyu Song, Wei-Nan Zhang, Jingwen Hu, and Ting Liu. 2020. Generating Persona Consistent Dialogues by Exploiting
Natural Language Inference. Proceedings of the AAAI Conference on Artificial Intelligence 34, 05 (Apr. 2020), 8878–8885.
https://doi.org/10.1609/aaai.v34i05.6417
[170] Kaiqiang Song, Logan Lebanoff, Qipeng Guo, Xipeng Qiu, Xiangyang Xue, Chen Li, Dong Yu, and Fei Liu. 2020. Joint
Parsing and Generation for Abstractive Summarization. In Proceedings of the AAAI Conference on Artificial Intelligence,
Vol. 34. 8894–8901.
[171] Kai Song, Yue Zhang, Heng Yu, Weihua Luo, Kun Wang, and Min Zhang. 2019. Code-Switching for Enhancing NMT
with Pre-Specified Translation. NAACL HLT 2019 - 2019 Conference of the North American Chapter of the Association
for Computational Linguistics: Human Language Technologies - Proceedings of the Conference 1 (4 2019), 449–459.
https://arxiv.org/abs/1904.09107v4 | SurveyofHallucinationinNatural Language Generation |
Wikipedia | Tool Learning with Foundation Models |
Hinton et al. (2015) proposed network distillation as a way to transfer the knowledge from an
ensemble of many separately-trained networks into a single, typically compact network, performing
a type of model compression. In this paper, we are considering a related but orthogonal task: rather
than distilling the model, we propose to distill the dataset. Unlike network distillation, we keep the
model fixed but encapsulate the knowledge of the entire training dataset, which typically contains
thousands to millions of images, into a small number of synthetic training images. We show that we
can go as low as one synthetic image per category, training the same model to reach surprisingly
good performance on these synthetic images. For example in Figure 1a, we compress 60, 000 training
images of MNIST digit dataset into only 10 synthetic images (one per class), given a fixed network
initialization. Training the standard LENET (LeCun et al., 1998) on these 10 images yields test-time | DATASET DISTILLATION |
Evaluations Workstream
Gaurav Mishra, Co-Lead
Jonathan H. Clark, Co-Lead
Mark Omernick, Co-Lead
Sebastian Ruder, Co-Lead (Tech Report)
Melvin Johnson, Core Contributor
Yanping Huang, Core Contributor
Ambrose Slone, Contributor
Andrea Hu, Contributor
Andrew M. Dai, Contributor
Colin Cherry, Contributor
Denny Zhou, Contributor
Gustavo Hernandez Abrego, Contributor
Jacob Austin, Contributor
Jan Botha, Contributor
John Wieting, Contributor
Joshua Maynez, Contributor
Kathleen Kenealy, Contributor
Kefan Xiao, Contributor
Kelvin Xu, Contributor
Kevin Brooks, Contributor
Linting Xue, Contributor
Markus Freitag, Contributor
Martin Polacek, Contributor
Maxim Krikun, Contributor
Michele Catasta, Contributor
Orhan Firat, Contributor
Parker Riley, Contributor
Pengcheng Yin, Contributor
Sebastian Gehrmann, Contributor
Siamak Shakeri, Contributor
Xavier Garcia, Contributor
Xuezhi Wang, Contributor | PaLM 2 Technical Report |
our Workspace updates
. And just like with Smart
Since the early days of Street View, AI has stitched together billions of panoramic images, so people can explore
the world from their device. At last year’s I/O we introduced Immersive View, which uses AI to create a high-fidelity
representation of a place, so you can experience it before you visit.
Now, we’re expanding that same technology to do what Maps does best: help you get where you want to go.
Google Maps provides 20 billion kilometers of directions, every day — that’s a lot of trips. Now imagine if you could
see your whole trip in advance. With
you can, whether you're walking, cycling or driving.
Immersive View for routes | Google I_O 2023_ Making AI more helpful for everyone |
Ondřej Dušek, David M Howcroft, and Verena Rieser. Semantic Noise Matters for Neural Natural Language
Generation. In Proceedings of the 12th International Conference on Natural Language Generation (INLG 2019),
pp. 421–426, Tokyo, Japan, 2019. URL https://www.aclweb.org/anthology/W19-8652/.
Julian Martin Eisenschlos, Maharshi Gor, Thomas Müller, and William W Cohen. Mate: Multi-view attention
for table transformer efficiency. arXiv preprint arXiv:2109.04312, 2021.
Alexander R Fabbri, Irene Li, Tianwei She, Suyi Li, and Dragomir R Radev. Multi-news: A large-scale
multi-document summarization dataset and abstractive hierarchical model. arXiv preprint arXiv:1906.01749,
2019.
William Fedus, Barret Zoph, and Noam Shazeer. Switch transformers: Scaling to trillion parameter models
with simple and efficient sparsity. arXiv preprint arXiv:2101.03961, 2021.
Wikimedia Foundation. Acl 2019 fourth conference on machine translation (wmt19), shared task: Machine | UL2- Unifying Language Learning Paradigms |
B) Uncertain Event Sequences: arise from a number of sources including measurement error,
randomness in the underlying phenomenon, and due to distributed and asynchronous data
gathering. They are used in a number of real-world scenarios to model and analyse spatial or
temporal data, which is of interest in diverse disciplines as computational neuroscience, earth
science and telecommunications. Marked event sequences are even more general and can be
applied to computer and economic systems for examples.
Uncertain Time Series: are most naturally associated with measurement errors, but can
directly represent a range of variation (e.g. high/low stock prices in a day's trading,
confidence intervals for predictions) or deliberate obfuscation for reasons of privacy
preservation. They can be seen as special cases of event sequences, but while in uncertain
time series the uncertainty lies in the value, in uncertain event sequences, the uncertainty is in | informatics-phd-projects-2022-23 |
C.4. Additional results for filtering and clustering | alphacode |
The range of failures in language understanding of current Transformers such as GPT-2
(see Marcus, 2019, 2020) reflects something similar: the schism between predicting
general tendencies (like the likelihood of the phrase mom's house appearing the
neighborhood of the words and phrases such as drop, off, pick, up and clothing in the
corpora GPT-2) and the capacity to represent, update, and manipulate cognitive
models. When BERT and GPT-2 failed to track where the dry cleaning would be it was a
direct reflection of the fact GPT and BERT have no representation of the properties of individual
entities as they evolve over time. Without cognitive models, systems like these are lost.
Sometimes they get lucky from statistics, but lacking cognitive models they have no
reliable foundation with which to reason over. | The Next Decade in AI- |
https://doi.org/10.1017/9781108890960 Published online by Cambridge University Press
102
Samuel C. Woolley
computational propaganda, and, correspondingly, disinformation and online
polarization. Quantitative insight
into the roles of automation, network
structure, temporal markers, and message semantics over social media can allow
experienced researchers to effectively create ways of measuring the flow of political
manipulation over social media over sustained periods. The results of longitudinal
research on this phenomenon will be crucial to building evolving long-term public
and governmental understandings of computational propaganda. | Social_Media_and_Democracy |
Temp → Learn 0.05 0.07 0.2 1.0
SUN-D 24.1 27.0 27.3 26.7 28.0
54.8 56.7 52.4 45.4 24.3
ESC
(a) Temperature for loss.
Spatial align → None Aligned
26.7
(e) Spatial alignment of depth.
SUN-D
16.0
Proj head → Linear MLP
26.7 26.5
56.7 51.0
SUN-D
ESC
(b) Projection Head.
Data aug → None RandErase
SUN-D
24.2
26.7
(f) Depth data aug.
32
Epochs → 16
64
SUN-D 26.7 27.9 29.9
56.7 61.3 62.9
ESC
(c) Training epochs.
Temporal align→ None Aligned
56.7
ESC
55.7
(g) Temporal alignment of audio.
SUN-D
Data aug → Basic Strong
26.7
22.6
(d) Data aug for image.
25.4
56.7
ESC
Data aug → Basic +Freq mask
56.5
ESC
(h) Audio data aug.
56.7 | IMAGEBIND- One Embedding Space To Bind Them A |
L
i
m
i
t
e
d
h
y
p
o
t
h
e
s
i
s
s
p
a
c
e
T
o
u
n
d
e
r
s
t
a
n
d
t
r
a
n
s
f
o
r
m
e
r
m
o
d
e
l
s
m
o
r
e
f
u
l
l
y
w
e
w
i
l
l
n
e
e
d
t
o
m
o
v
e
f
r
o
m
i
n
t
e
r
p
r
e
t
i
n
g
s
i
n
g
l
e
n
e
u
r
o
n
s
t
o
i
n
t
e
r
p
r
e
t
i
n
g
c
i
r
c
u
i
t
s
.
T
h
i
s
w
o
u
l
d
m
e
a
n
i
n
c
l
u
d
i
n
g
h
y
p
o
t
h
e
s
e
s
a
b
o
u
t
d
o
w
n
s
t
r
e
a
m
e
f
f
e
c
t
s
o
f
n
e
u
r
o
n
s
,
h
y
p
o
t
h
e
s
e
s
a
b
o
u
t
a
t
t
e
n
t
i
o
n
h
e
a
d
s
a
n
d
l
o
g
i
t
s
,
a
n
d
h
y
p
o
t
h
e
s
e
s
i
n
v
o
l
v
i
n
g
m
u
l
t
i
p
l
e
i
n
p
u
t
s
a
n
d
o
u
t
p
u
t
s
.
E
v
e
n
t
u
a
l
l
y
,
o
u
r
e
x
p
l
a
i
n
e
r
m
o
d
e
l
s
w
o
u
l
d
d
r
a
w
f
r
o
m
a
r
i
c
h
s
p
a
c
e
o
f
h
y
p
o
t
h
e
s
e
s
,
j
u
s
t
l
i
k
e
i
n
t
e
r
p
r
e
t
a
b
i
l
i
t
y
r
e
s
e
a
r
c
h
e
r
s
d
o
.
C
o
m
p
u
t
a
t
i
o
n
a
l
r
e
q
u
i
r
e
m
e
n
t
s
O
u
r
m
e
t
h
o
d
o
l
o
g
y
i
s
q
u
i
t
e
c
o
m
p
u
t
e
-
i
n
t | Language models can explain neurons in language models |
potentially affect LLM developers who gather vast amounts of public data from the internet, which
may include personal information. Obtaining explicit consent from data creators is difficult at this
scale, and it is uncertain whether other legal grounds exist for processing this personal information.
Moreover, even with a valid legal basis, GDPR mandates that data processors inform individuals as
to how their data is being processed and provide data access controls, such as the right to have your
data deleted or to modify erroneous data. This would require LLM providers to be transparent about
the data they have collected and provide tooling for individuals to inspect their data and have the
possibility to delete it.
The lack of transparency and openness surrounding the development processes of generative AI
models has also raised concerns in the scientific community. Some of the best-performing LLMs, | StarCoder_paper (1) |
top-K tokens.
We presented a detailed study about how different routing decisions affect the instruct fine-tuning
performance in Figure 3 and Table 1, which includes the checkpoints from Switch Transformer
top-1 token-choice gating (FLAN-Switch), GShard top-2 token-choice gating (FLAN-GS) and expert-
choice top-2 gating (FLAN-EC) models pre-trained on the same GLaM [10] dataset. It is evident
that activating more experts, as demonstrated by the comparison between the FLAN-Switch and
FLAN-GS strategies, results in enhanced performance across all four benchmarks. Among these
benchmarks, the MMLU-Direct model shows the most significant improvement, with an increase
from 38.0% to 39.9% for BASE/LARGE-sized models. Although the gains at the extra-large scale
are more modest, they remain noteworthy and meaningful. It’s noteworthy that instruction-tuning
significantly amplifies the performance of both held-out MMLU, BBH, and held-in QA and reasoning | Mixture-of-Experts |
In the final set of items of SHAPE , there is no explicit reference to privacy threats, which is interesting given
that only one item related to privacy was removed, while the remaining items underwent filtering in the EFA. The
absence of explicit representation of privacy concerns among the filtered items may suggest that we considered
them to be less relevant compared to other factors, such as the agency of the augmented human or the perceived
threat it poses to the observer. Furthermore, we acknowledge that the subscale "Social Threat" may not specifically
target any particular type of threat, including privacy threats. Therefore, it is possible that this subscale captures
certain aspects of privacy concerns, even in the absence of explicit references to privacy threats. | Society’sAttitudesTowardsHumanAugmentation |
experienced bees estimate the sun’s course. Proceedings of the National Academy of Sciences, 91(10), 4471-
4474.
Engelcke, M., Kosiorek, A. R., Jones, O. P., & Posner, I. (2019). GENESIS: Generative Scene Inference and
Sampling with Object-Centric Latent Representations. arXiv, 1907.13052v3.
Evans, R., & Grefenstette, E. (2017). Learning Explanatory Rules from Noisy Data. arXiv, cs.NE.
Fawzi, A., Malinowski, M., Fawzi, H., & Fawzi, O. (2019). Learning dynamic polynomial proofs. cs.LG.
Fodor, J. A., & Pylyshyn, Z. W. (1988). Connectionism and cognitive architecture: a critical analysis.
Frankland, S. M., & Greene JD. (2019). Concepts and Compositionality: In Search of the Brain’s Language
Cognition, 28(1-2), 3-71.
of Thought. Annual review of psychology.
Gallistel, C. R. (1990). The organization of learning. The MIT Press.
Gallistel, C. R., & King, A. P. (2010). Memory and the computational brain: Why cognitive science will transform
neuroscience. John Wiley & Sons. | The Next Decade in AI- |
$10 x 3 = $30
So the total amount Lee earned was:
$528 + $30 = $558
However Guanaco can breakdown at even simple problems if it fails to break them down step-by-step,
a known issue [63]. For example, we see the following interaction:
User: Please factorize 1833.
Guanaco:
which means it has only two factors:
Sure, I’d be happy to help!
So the unique factorization of 1833 is:
1833 is a prime number,
1 and 1833.
2ˆ1 * 3ˆ2 * 17ˆ1
I hope this helps!
questions.
Let me know if you have any other
which is not only wrong (the true factorization is 3 × 17 × 43), it’s wrong twice .
Theory of Mind Guanaco displays surprisingly strong Theory of Mind capabilities [41, 51]. For
instance, the model correctly answers the following question in great detail: | QLORA |
Categorical Modelling. The class labels in AudioSet can be arranged hierarchically to obtain the
following top-level categories: i) Human sounds, ii) Animal sounds, iii) Natural sounds, iv) Sounds
of Things, v) Channel, environment, background sounds, vi) Source-ambiguous sounds, and vii)
Music. We map the class labels in AudioCaps to the seven main categories listed above. The Music
category is very rare in AudioCaps and the rest either appear on their own or in various combinations
with others. We select the most frequently occurring category combinations and analyze the perfor-
mance of various models within the constituting AudioCaps instances in Table 5. The performance
of the two models is pretty balanced across the FD and KL metrics, with TANGO being better in
8 | Text-to-Audio Generation using Instruction-Tuned LLM and Latent Diffusion Model |
n
e
w
a
v
e
n
u
e
f
o
r
A
I
-
a
s
s
i
s
t
e
d
t
e
x
t
g
e
n
e
r
a
t
i
o
n
o
n
c
u
r
r
e
n
t
a
f
f
a
i
r
s
.
J
u
r
a
s
s
i
c
-
X
c
a
n
a
s
s
i
s
t
i
n
t
e
x
t
g
e
n
e
r
a
t
i
o
n
o
n
u
p
-
t
o
-
d
a
t
e
e
v
e
n
t
s
b
y
c
o
m
b
i
n
i
n
g
a
p
o
w
e
r
f
u
l
l
a
n
g
u
a
g
e
m
o
d
e
l
w
i
t
h
a
c
c
e
s
s
t
o
W
i
k
i
d
a
t
a
P
e
r
f
o
r
m
i
n
g
m
a
t
h
o
p
e
r
a
t
i
o
n
s
A
6
y
e
a
r
o
l
d
c
h
i
l
d
l
e
a
r
n
s
m
a
t
h
f
r
o
m
r
u
l
e
s
,
n
o
t
o
n
l
y
b
y
m
e
m
o
r
i
z
i
n
g
e
x
a
m
p
l
e
s
.
I
n
c
o
n
t
r
a
s
t
,
l
a
n
g
u
a
g
e
m
o
d
e
l
s
a
r
e
d
e
s
i
g
n
e
d
t
o
l
e
a
r
n
f
r
o
m
e
x
a
m
p
l
e
s
,
a
n
d
c
o
n
s
e
q
u
e
n
t
l
y
a
r
e
a
b
l
e
t
o
s
o
l
v
e
v
e
r
y
b
a
s
i
c
m
a
t
h
l
i
k
e
1
| Jurassic-X_ Crossing the neuro-symbolic chasm with the MRKL system |
Review,
plagues-mexicos-election/
Paavola, J., Helo, T., Jalonen, H., Sartonen, M., & Huhtinen, A.-M.
(2016).
Understanding the trolling phenomenon: The automated detection of bots and
cyborgs in the social media. Journal of Information Warfare, 15(4), 100–111.
Ratkiewicz, J., Conover, M., Meiss, M., Goncalves, B., Flammini, A., & Menczer, F.
(2011). Detecting and tracking political abuse in social media. In Proceedings of the
Fifth International AAAI Conference on Weblogs and Social Media (ICWSM).
Barcelona: AAAI Press. www.aaai.org/ocs/index.php/ICWSM/ICWSM11/paper/
viewFile/2850/3274
Ratkiewicz, J., Conover, M., Meiss, M. et al. (2011). Truthy: Mapping the spread of
astroturf in microblog streams. In Proceedings of the 20th International Conference
Companion on World Wide Web (pp. 249–252). Hyderabad: ACM. http://dl
.acm.org/citation.cfm?id=1963301
Robb, J. (2007). When bots attack. Wired, August 23. www.wired.com/2007/08/ff-
estonia-bots/ | Social_Media_and_Democracy |
Gemini Ultra achieves state-of-the-art results on various few-shot video captioning tasks as well as
zero-shot video question answering tasks as shown in Table 10. This demonstrates its capability of
strong temporal reasoning across several frames. Figure 21 in the appendix provides a qualitative
example of understanding the video of the ball-striking mechanics of a soccer player and reasoning
about the player can improve their game.
Gemini Ultra
62.7
4-shots
Gemini Pro
57.4
4-shots
Few-shot SoTA
56.0
DeepMind Flamingo, 4-shots
Task
VATEX (test)
English video captioning
(Wang et al., 2019)
VATEX ZH (test)
Chinese video captioning
(Wang et al., 2019)
YouCook2 (val)
English cooking video captioning
(Zhou et al., 2018)
NextQA (test)
Video question answering
(Xiao et al., 2021)
ActivityNet-QA (test)
Video question answering
(Yu et al., 2019)
Perception Test MCQA (test)
Video question answering
(Pătrăucean et al., 2023)
51.3
4-shots
135.4
4-shots
29.9
0-shot
52.2
0-shot
54.7
0-shot | gemini_1_report |
Long generation and story mode.
In MusicLM, gene-
ration is autoregressive in the temporal dimension which
makes it possible to generate sequences longer than those
used during training. In practice, the semantic modeling
stage is trained on sequences of 30 seconds. To generate
longer sequences, we advance with a stride of 15 seconds,
using 15 seconds as prefix to generate an additional 15 sec-
onds, always conditioning on the same text description.
With this approach we can generate long audio sequences
which are coherent over several minutes.
With a small modification, we can generate long audio se-
quences while changing the text description over time. Bor-
rowing from Villegas et al. (2022) in the context of video
generation, we refer to this approach as story mode. Con-
0.01.02.05.010.0Prompt length [sec]0.1%1%10%Fraction of matching examplesapproximate matchesexact matchesMusicLM: Generating Music From Text | MusicLM |
Survey of Hallucination in Natural Language Generation
17
Controllability. Controllability means the ability of models to control the level of hallucination
and strike a balance between faithfulness and diversity [41, 159]. As mentioned in Section 3, it is
acceptable for chit-chat models to generate a certain level of hallucinatory content as long as it is
factual. Meanwhile, for the abstractive summarization task, there is no agreement in the research
community about whether factual hallucinations are desirable or not [125]. Therefore, we believe
controllability merits attention when exploring hallucination mitigation methods. | SurveyofHallucinationinNatural Language Generation |
volving objects that were unseen in either the original robot
dataset or the finetuning datasets, e.g. a toy turtle (Fig. 5, d). | PaLM-E- An Embodied Multimodal Language Model |
Delgado, R. (1982). Words that wound: A tort action for racial insults, epithets, and
name calling. Harvard Civil Rights-Civil Liberties Review, 17, 133–181.
Delgado, R., & Stefancic, J. (2014). Hate speech in cyberspace. Wake Forest Law
Review, 49, 319. https://ssrn.com/abstract=2517406
DellaVigna, S., Enikolopov, R., Mironova, V., Petrova, M., & Zhuravskaya, E. (2014).
Cross-border media and nationalism: Evidence from Serbian radio in Croatia.
American Economic Journal: Applied Economics, 6(3), 103–132.
Dinakar, K., Reichart, R., & Lieberman, H. (2011). Modeling the detection of Textual
Cyberbullying. The Social Mobile Web, 11(02), 11–17.
Djuric, N., Zhou, J., Morris, R., Grbovic, M., Radosavljevic, V., & Bhamidipati, N.
(2015). Hate speech detection with comment embeddings. In A. N. Joinson,
K. Y. A. McKenna, T. Postmes, & U.-D. Reips (Eds.), Proceedings of the 24th
International Conference on World Wide Web (pp. 29–30). New York: ACM. | Social_Media_and_Democracy |
13
Table 5: Evaluation on reasoning tasks. We show the number of exemplars in brackets. PaLM 2 results are using its
instruction-tuned variant (see Appendix A.2) except for XCOPA; PaLM 2 results on ARC-C, StrategyQA, and CSQA
use chain-of-thought prompting (CoT; Wei et al., 2022) and self-consistency (SC; Wang et al., 2023). PaLM 2 results
on BB Hard use CoT. Superscripts denote results from past work: aGPT-4 (OpenAI, 2023b), bPaLM (Chowdhery et al.,
2022), cPaLM+CoT+SC (Wang et al., 2023), dQDGAT (Chen et al., 2020), eDeBERTaV3-large+KEAR (Xu et al.,
2022), f PaLM+CoT (Suzgun et al., 2022), gPaLM+CoT (Shi et al., 2023).
WinoGrande
ARC-C
DROP
StrategyQA
CSQA
XCOPA
BB Hard
(5)
(25)
(3)
SOTA GPT-4
87.5a
87.5a
96.3a
96.3a
80.9a
88.4d
81.6c
-
91.2e
-
89.9g
-
65.2 f
-
(5)
(4)
PaLM PaLM 2
85.1b
88.7c
70.8b
81.6c
80.7c
89.9g
65.2 f
90.9 (5)
95.1 (4)
85.0 (3)
90.4 (6)
90.4 (7)
94.4 (4)
78.1 (3)
(1)
(6)
(4)
(7)
(3) | PaLM 2 Technical Report |
LLMs typically suffer from arbitrary predictions—they
might produce invalid outputs (e.g., hallucination or invalid
formats)— which is detrimental to driving systems. To
investigate this effect, we conducted a stability test of our
Agent-Driver. Specifically, we used different amounts of
training data to instruct the LLMs in our system, and we
tested the number of invalid outputs during inference on the
validation set. As shown in Table 5, Agent-Driver exposed to
only 1% of the training data sees zero invalid output during
inference of 6,019 validation scenarios, suggesting that our
system attains high output stability with proper instructions.
3.7. Empirical Study
Effectiveness of system components. Table 3 shows the
results of ablating different components in Agent-Driver.
All variants utilize 10% training data for instructing the
LLMs. From ID 1 to ID 5, we ablate the main components in
Agent-Driver, respectively. We deactivate the self-reflection | ALanguageAgentforAutonomousDriving |
3.1 UNIQUE VIDEO GENERATION CAPABILITIES | IMAGEN VIDEO- HIGH DEFINITION VIDEO GENERATION WITH DIFFUSION MODELS |
Let the answer for q be denoted eans, and its
masked mention mans = (eans, sans, tans). For a
masked mention mans, define a query vector to
access the fact memory as:
vmans = WT
f [h(T )
sans ; h(T )
tans ]
(4)
sans and h(T )
where h(T )
tans are the contextual embeddings
for the start and end tokens of the mention mans,
and Wf is the linear transformation matrix into the
embedding space of head pairs A. | Adaptable and Interpretable Neural Memory Over Symbolic Knowledge |
In the next section, we discuss the details of this process.
3. Dataset Creation
One of the key reasons that there are not many generative music for video
systems out there, is the lack of symbolic music with video datasets. Given that
the accuracy of music transcription systems is constantly growing (Cheuk et al.,
2020, 2021, 2023), especially chord transcription (Park et al., 2019), we set out
11
to design a novel way to create a dataset. This resulted in a new dataset, called
MuVi-Sync, comprising both music features and video features extracted from
a total of 748 music videos. Below, we describe the music and video features
that we extract from this dataset.
3.1. Music Features
From the audio track of the music video, we extract four essential features:
note density, loudness, chords, and key. These features play a crucial role in
capturing the musical characteristics and composition of the audio. The chord | Video2Music |
are seen as selfish and shallow, only interested in high-status and physically attractive men, while completely ignoring men who are perceived as less attractive. According to incels, women are unempathetic towards their struggles and contribute to the unfairness of the dating game.“Jailbreak” PromptGPT-4 (launch)Attack Typesafe usage. | gpt-4-system-card |
In the following section, we provide a brief overview of the literature on
misinformation in political science and psychology, which provides a basis for
understanding the phenomena discussed in this chapter. We then turn to what
we know about the production of disinformation and the supply and
availability of misinformation more broadly online. We then focus on the
consumption side, with a section on exposure and its correlates on the
individual
factor determining exposure is how
misinformation is spread and disseminated, which we cover next. The
penultimate
effects of
misinformation and how it is studied. We conclude with a discussion of gaps
in our knowledge and future directions in research in this area. | Social_Media_and_Democracy |
research questions remain regarding SSL’s generalization guarantees, fairness proper-
ties, and robustness to adversarial attacks or even naturally occurring variations. Such
questions are crucial to the reliability of SSL methods. | A Cookbook of Self-Supervised Learning |
Theorem 20.
(1) P2↓ (cid:3) PL↓,
(2) PL↓ (cid:3) P2↓,
(3) P2↑ (cid:3) PL↑,
(4) PL↑ (cid:3) P2↑.
Proof. (1) Remove the path from s01 to s11 in Fig. 2(b). The result is P2↓ but not PL↓ since the path t0, t1, t2, t3 is not
loosely downwards state refinable.
(3–4) Analogous to (1–2). (cid:2)
(2) Example 17(2) is PL↓ but not P2↓.
The two types of hierarchies are illustrated for the downwards direction in Fig. 4 (note that PW↓ ⇒ Pk↓ for all k by
definition), where the arrows denote the ⇒ and (cid:3) relationships between the properties. All these hierarchies collapse if
the transformation also has property M in the same direction.
Theorem 21. (1) M↓P1↓ ⇒ PS↓,
(2) M↑P1↑ ⇒ PS↑.
14
C. Bäckström and P. Jonsson
Artificial Intelligence 302 (2022) 103608 | A-framework-for-analysing-state-abstraction-metho_2022_Artificial-Intelligen |
3) ARTISTS
The WikiArt dataset includes artworks by more than 2000 dif-
ferent artist, represented with a varying number of images.
For the purpose of our exploration, we choose a subset
of 20 well known artists, belonging to different historical art
movements. Box plots in Fig. 12 show the distribution of the
predicted scores for different artists.
The arbitrary choice of artists hinders us from making any
general conclusions, however the relative relations between
the predicted aesthetic, sentiment and memorability scores
of the chosen artists still yield interesting outcomes. For
instance, the case of William Turner whose works have the | A_Deep_Learning_Perspective_on_Beauty_Sentiment_and_Remembrance_of_Art |
0100200300400500600700800900100011001200Diffusion Timestep (t)0102030405060708090100Normalized Edit DistancePythonCF RuleBash7 Limitations
CODEFUSION is not a global system as we only
consider natural language utterances in English.
Furthermore, natural language specifications can
be provided at varying levels of detail – utterances
with less detail may result in worse performance.
We consider various programming languages, but
more complex languages may result in worse per-
formance. We also find that CODEFUSION strug-
gles when tasked with generating longer code snip-
pets or programs that have long-ranged syntactic
dependencies. Because diffusion-based models de-
noise iteratively, inference latency is substantial,
rising exponentially with target generation length. | CODEFUSION |
Online Political Advertising in the United States
135
negative. Although the impact of internet ads was smaller than for television,
when one considers cost, the return on investment was just as high. More
research like this is needed to get a full assessment of the impact of digital
advertising, but other studies have found no impact. One experiment targeted
legislators’ constituents with Facebook ads but found no difference in
people’s evaluations or recognition of
the candidate shown in the ad
between those who saw the ad repeatedly and those who never saw it
(Broockman and Green 2014). Another experiment found no impact of
exposure to a week’s worth of online display ads on people’s views toward
the Black Lives Matter movement (Coppock and Broockman 2015). One
study did reveal a small impact of exposure to banner and pre-roll ads on
voter turnout
in a municipal election – but only when the race was
competitive (Haenschen and Jennings 2019). | Social_Media_and_Democracy |
energy consumption, memory usage, and processing power. Establishing these bench-
marks is essential for advancing the development of more resource-efficient LLMs, a
key priority given the increasing size and complexity of these models. | Beyond Efficiency |
inpainting module, PixelSynth is prone to generating incoher-
ent and blurry content, especially in the inpainted regions.
Moreover, as shown in Fig. 6, our Text2NeRF supports text-
driven scene generation in a large view range thanks to
our progressive scene inpainting and updating strategy. On
the other hand, other novel view synthesis methods produce
blurred scene-filling results even at a small viewing angle since
the text-related guidance is not considered in such methods. | Text2NeRF- Text-Driven 3D Scene Generation with Neural Radiance Fields |
LocalFoilTrees[55],orLoRE[56].Theseapproachesbuildsurrogate
models for each prediction sample, learning the reference model’s
behaviorontheparticularcaseofinterestbyintroducingperturbations
to the feature vector variables. By doing so, they can provide a local
featureimportanceestimate,whichisconsideredanindirectmethodto
explainamodel[32].Inparticular,weoptedforthefeaturerelevance
explanationaboveotherpost-hocmodelagnostictechniques(e.g.,rule
extraction) due to the simplicity of implementation. While other ex-
planationscanalsobebuiltconsideringtheontologyabstraction,care
mustbeputtoensuretheseremainvalid.E.g.,inthecaseofruleextrac-
tion,thenon-contradictionprinciplemustbeenforcedsincewhilethe
rulecanbevalidatafeaturelevel,thesameisnotguaranteedathigher
abstractionlevels.Whilesuchapproachesdonotconsiderthetemporal
dimensionperse,features’metadatacanbeconsideredtocaptureitand
enrichtheforecastexplanations[57].
When designing a system providing explanations for AI models, | Knowledge-graph-based-rich-and-confidentiality-preserving-Ex_2022_Informatio |
classification since BiomedGPT can fully understand it from the pretraining stage. However, addressing
task-aware instruction sensitivity remains an open question. | BiomedGPT |
more qualitative results that demonstrate how different amounts of safety data in training can change model
behavior in responding to adversarial and non-adversarial prompts. | Llama2 |
to violent clashes and undermining social cohesion (Izsak 2015). For example,
Facebook has come under fire for its role in mobilizing anti-Muslim mob
violence in Sri Lanka and for inciting violence against the Rohingya people in
Myanmar (Vindu, Kumar, and Frenkel 2018). Elucidating the mechanisms by
which exposure to hate speech drives intergroup tension, survey data and
experimental evidence from Poland suggest that frequent and repetitive
exposure to hate speech leads to desensitization to hateful content, lower
evaluations of populations targeted by hate speech, and greater distancing –
resulting in higher levels of anti–out-group prejudice (Soral et al. 2018). | Social_Media_and_Democracy |
being called in a loop or recursively, thus leading to a number of subtle bugs (see, for example,
the exploit of the DAO bug, leading to loss of $50 Millions). It is then reasonable to expect
that the best way to formally verify smart contracts is by using modular reasoning: for each
smart contract, the other contracts with which it interacts can be considered an environment.
This environment can be overapproximated using learning techniques in combination with
sampling and traditional model checking approaches. After verification of a single contract
passes successfully, some symbolic representations of the contracts with respect to the
correctness properties will be combined to prove correctness of the overall system.
The project will include a significant implementation component. The implementation is
done using the software verification platform developed at USI. The main development task | informatics-phd-projects-2022-23 |
QUESTION: Can you bring me something salty?
MODEL ANSWER (INCORRECT): Explanation: The user would like something salty. There are several
options in the objects list, so I will bring the user a lime soda. Plan: find(lime soda), pick(lime soda), find(user),
put(lime soda). (cid:55) (Feasible answers including finding chips. The model misidentify lime soda as salty.) | Chain-of-Thought Prompting Elicits Reasoning in Large Language Models |
Yuntao Bai, Saurav Kadavath, Sandipan Kundu, Amanda Askell, Jackson Kernion, Andy Jones,
Anna Chen, Anna Goldie, Azalia Mirhoseini, Cameron McKinnon, Carol Chen, Catherine Ols-
son, Christopher Olah, Danny Hernandez, Dawn Drain, Deep Ganguli, Dustin Li, Eli Tran-
Johnson, Ethan Perez, Jamie Kerr, Jared Mueller, Jeffrey Ladish, Joshua Landau, Kamal Ndousse,
Kamile Lukosuite, Liane Lovitt, Michael Sellitto, Nelson Elhage, Nicholas Schiefer, Noemi Mer-
cado, Nova DasSarma, Robert Lasenby, Robin Larson, Sam Ringer, Scott Johnston, Shauna
Kravec, Sheer El Showk, Stanislav Fort, Tamera Lanham, Timothy Telleen-Lawton, Tom Con-
erly, Tom Henighan, Tristan Hume, Samuel R. Bowman, Zac Hatfield-Dodds, Ben Mann, Dario
Amodei, Nicholas Joseph, Sam McCandlish, Tom Brown, and Jared Kaplan. Constitutional AI:
Harmlessness from AI Feedback, December 2022. URL http://arxiv.org/abs/2212.
08073. arXiv:2212.08073 [cs]. | CHAIN-OF-THOUGHTREASONING IS APOLICY IMPROVEMENTOPERATOR |
Alon Talmor, Jonathan Herzig, Nicholas Lourie, and
Jonathan Berant. 2019. CommonsenseQA: A ques-
tion answering challenge targeting commonsense
knowledge. In Proceedings of the 2019 Conference
of the North American Chapter of the Association for
Computational Linguistics: Human Language Tech-
nologies, Volume 1 (Long and Short Papers), pages
4149–4158, Minneapolis, Minnesota. Association for
Computational Linguistics.
Zheng Tang, Gus Hahn-Powell, and Mihai Surdeanu.
2020. Exploring interpretability in event extraction:
Multitask learning of a neural event classifier and
an explanation decoder. In Proceedings of the 58th
Annual Meeting of the Association for Computational
Linguistics: Student Research Workshop, pages 169–
175, Online. Association for Computational Linguis-
tics.
Ian Tenney, Dipanjan Das, and Ellie Pavlick. 2019.
In
BERT rediscovers the classical NLP pipeline. | Measuring Association Between Labels and Free-Text Rationales |
Human Evaluation. We also conduct human evaluation on the general quality of the model
responses on the combined test set described in subsection 3.1, which covers several existing
2The specific version of the data we used is https://huggingface.co/datasets/WizardLM/
WizardLM_evol_instruct_V2_196k/tree/main.
7
102103104Data Size010203040506070Win RateHumpbackWizardLLMAlpaca GPT-4Vicuna (sharegpt)OALIMAAlpacaFLAN v2Non-distillDistilledFigure 5: Scaling up self-curated instruction data A5 brings improvement in both small (7B) and
large (65B) LLaMa finetuned models, and neither model is saturated with 40,000 instructions.
Labelled Examples Win Rate %
3k
9k
1k
3k
Non-
distilled
65B
Non-
distilled
33B
Distilled
Proprietary
Humpback 65B
Guanaco 65B
LIMA 65B
Humpback 33B
OASST RLHF 33B 161k
9k
Guanaco 33B
161k
OASST SFT 33B
140k
Vicuna 33B
190k
WizardLLM 13B
airoboros 65B
17k
Falcon Instruct 40B 100k
GPT-4
Claude 2
ChatGPT
Claude | Self-AlignmentwithInstructionBacktranslation |
Shahaf Bassan, Yossi Adi, and Jeffrey S Rosenschein. Unsupervised symbolic music segmentation
using ensemble temporal prediction errors. arXiv preprint arXiv:2207.00760, 2022.
Adrien Ycart, Emmanouil Benetos, et al. A study on lstm networks for polyphonic music sequence
modelling. ISMIR, 2017.
Shulei Ji, Jing Luo, and Xinyu Yang. A comprehensive survey on deep music generation: Multi-level
representations, algorithms, evaluations, and future directions. arXiv preprint arXiv:2011.06801,
2020.
Prafulla Dhariwal, Heewoo Jun, Christine Payne, Jong Wook Kim, Alec Radford, and Ilya Sutskever.
Jukebox: A generative model for music. arXiv preprint arXiv:2005.00341, 2020.
Chuang Gan, Deng Huang, Peihao Chen, Joshua B Tenenbaum, and Antonio Torralba. Foley
music: Learning to generate music from videos. In Computer Vision–ECCV 2020: 16th European
Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XI 16, pages 758–775. Springer,
2020. | Simple and Controllable Music Generation |
[69] Zehan Wang, Haifeng Huang, Yang Zhao, Ziang Zhang,
and Zhou Zhao. Chat-3D: Data-efficiently Tuning Large
Language Model for Universal Dialogue of 3D Scenes.
arXiv preprint arXiv:2308.08769, 2023. 2
[70] Ho-Hsiang Wu, Prem Seetharaman, Kundan Kumar, and
Juan Pablo Bello. Wav2CLIP: Learning Robust Audio
In ICASSP, pages 4563–
Representations From CLIP.
4567. IEEE, 2022. 4
[71] Shengqiong Wu, Hao Fei, Leigang Qu, Wei Ji, and Tat-
Seng Chua. NExT-GPT: Any-to-Any Multimodal LLM.
arXiv preprint arXiv:2309.05519, 2023. 2, 3, 5, 7
[72] Yusong Wu, Ke Chen, Tianyu Zhang, Yuchen Hui, Taylor
Berg-Kirkpatrick, and Shlomo Dubnov. Large-scale Con-
trastive Language-Audio Pretraining with Feature Fusion
and Keyword-to-Caption Augmentation. In IEEE Interna-
tional Conference on Acoustics, Speech and Signal Pro-
cessing, ICASSP, 2023. 7, 14 | M2UGen |
Lichang Chen, Shiyang Li, Jun Yan, Hai Wang, Kalpa Gunaratna, Vikas Yadav, Zheng Tang, Vijay
Srinivasan, Tianyi Zhou, Heng Huang, et al. Alpagasus: Training a better alpaca with fewer data.
arXiv preprint arXiv:2307.08701, 2023c.
Lingjiao Chen, Matei Zaharia, and James Zou. How is chatgpt’s behavior changing over time? arXiv
preprint arXiv:2307.09009, 2023d. | ChatGPT’sOne-yearAnniversary-AreOpen-Source LargeLanguageModelsCatchingup |
5.3 Real Speech Data
For real speech experiment, we used Common Voice 11. Natural (non-synthesized) monolingual
speech-text datasets both in English and Spanish were used for training. The evaluation of Spanish-
English real speech Translation was conducted using real speech with verified translation from
the CoVoST2 test set Wang et al. [2020], which is a subset of the Common Voice 11 test set.
However, it should be noted there is no English-Spanish CoVoST2 test set or otherwise any test
set from Common Voice 11 with verified translation for English-Spanish and therefore only an
English-Spanish evaluation was omitted from this experiment. The proposed approach achieved an
10.67 in BLEU for the task of Spanish-English Translation which is an improvement of +0.75 in
BLEU over the baseline which achieves 9.92 BLEU. Audio samples are available in our website:
https://google-research.github.io/lingvo-lab/translatotron3.
5.4 Comparison to Supervised Approaches | Translatotron3 |
Recently, linear shape models dominate the representation
of statistical 3D model. Numerous methods [5, 12, 34, 41]
have shown PCA’s ability in modeling the human body and
face. Inspired by [37], we parameterize our character shape
linearly with the following equation,
MS = FS(B) = ¯MS +
βisi,
(2)
where ¯MS denotes the mean shape and MS is the recon-
structed shape. The coefficients of linear shape are βi ∈ B.
|B| is the number of shape parameters and is set to 100 in
our implementation. si ∈ R3×N denotes the orthogonal prin-
cipal components of vertex displacements that capture shape
variations in different character identities. The shape model
of RaBit is learned from 1,050 characters of 3DBiCar using
PCA [37]. RaBit’s eyeballs can be computed based on the
predefined landmarks shown in Fig. 4. Please refer to the
Supplementary for more details.
4.2. Pose Modeling | RaBit- Parametric Modeling of 3D Biped Cartoon Characters with a Topological-consistent Dataset |
put consistent across skeleton formats through output regu-
larization (Fig. 3c). We also experiment with direct latent
point prediction, and a hybrid variant for the last step. | Learning 3D Human Pose Estimation from Dozens of Datasets using a Geometry-Aware Autoencoder to Bridge Between Skeleton Formats |
• Multilinguality: We use 10 benchmarks: XLSum (Non-English languages) (Hasan et al., 2021),
WMT22 (Kocmi et al., 2022), WMT23 (Tom et al., 2023), FRMT (Riley et al., 2023), WikiLingua
(Non-English languages) (Ladhak et al., 2020), TydiQA (no context), TydiQA (GoldP) (Clark
et al., 2020), MGSM (Shi et al., 2023), translated MMLU (Hendrycks et al., 2021a), NTREX
(Federmann et al., 2022), FLORES-200 (Team et al., 2022). | gemini_1_report |
As there are many details that need to be specified in the above sketch to yield a very concrete model,
this gives rise to a wide range of interesting mechanism design challenges. Different properties of the
market require different mechanisms, where one can think of e.g. a static "one-shot" trading
scenario versus a scenario where agents can dynamically enter and exit the market, or indivisible
versus divisible goods, shareable vs unshareable goods, etc. In this project we will work on trying to
solve various challenging variants of this design problem. | informatics-phd-projects-2022-23 |
In parallel with the move from direct to distributed discovery, we have seen
the move to a digital media environment that affords people with more
opportunities for more participatory forms of news and media use, in the
process also exposing many to widespread online harassment and potentially
various forms of disinformation disseminated online and especially via
platforms. Digital media offers everyone with internet access a range of
both “web 1.0” and “web 2.0” ways of engaging in more participatory
forms of news and media use, ranging from commenting on news sites and
sharing via email to commenting and/or sharing via social media sites. While
all internet users have access to this participatory potential, it is important to
3 Beyond the specific issues of news diversity that we examine here, algorithmic and automated
ranking systems can embed various forms of discrimination and reinforce oppressive social
relations as, for example, Noble (2018) shows. | Social_Media_and_Democracy |
40000DemonAttack17.515.012.510.0DoubleDunk0250500750Enduro100500FishingDerby0102030Freeway100200300Frostbite02000040000Gopher250500750Gravitar10864IceHockey0200400600Jamesbond0500010000Kangaroo2000400060008000Krull02000040000KungFuMaster050100MontezumaRevenge100020003000MsPacman25005000750010000NameThisGame1000Pitfall20020Pong0500PrivateEye050001000015000Qbert25005000750010000Riverraid02000040000RoadRunner246Robotank050010001500Seaquest5001000SpaceInvaders02000040000StarGunner201510Tennis30004000TimePilot0100200300Tutankham0100000200000UpNDown040MFrames0510Venture040MFrames50000100000150000VideoPinball040MFrames20004000WizardOfWor040MFrames0200040006000ZaxxonA2CACERPPOFigure6:ComparisonofPPOandA2Conall49ATARIgamesincludedinOpenAIGymatthetimeofpublication.11A2CACERPPOAlien1141.71655.41850.3Amidar380.8827.6674.6Assault1562.94653.84971.9Asterix3176.36801.24532.5Asteroids1653.32389.32097.5Atlantis729265.31841376.02311815.0BankHeist1095.31177.51280.6BattleZone3080.08983.317366.7BeamRider30 | PPO |
and systematic process, enhancing the overall quality and fidelity of speech synthesis techniques. | AReviewofDeepLearningTechniquesforSpeechProcessing |
1Although there is a text preprocessing step in TTS systems,
We herein use preprocessed text interchangeably with the word
“text”.
Conditional Variational Autoencoder with Adversarial Learning for End-to-End Text-to-Speech | ConditionalVariationalAutoencoderwithAdversarialLearningfor End-to-EndText-to-Speech |
©2023 Cerebras Systems Inc. All Rights Reserved.
20
Cerebras-GPT: Open Compute-Optimal Language Models
1. HellaSwag is a dataset of multiple choice questions aimed to test a model’s common sense reasoning
abilities (Zellers et al., 2019). For example,
A woman is outside with a bucket and a dog. The dog is running around trying to avoid a
bath. She...
A. rinses the bucket off with soap and blow dry the dog’s head.
B. uses a hose to keep it from getting soapy.
C. gets the dog wet, then it runs away again.
D. gets into a bath tub with the dog.
The authors of the dataset adversely select examples such that they are difficult for language models
while still trivial for humans (with reported greater than 95% accuracy).
2. PIQA tests a model’s common sense reasoning about the physical world by posing a prompt and
two potential completions (Bisk et al., 2020). For example | Cerebras-GPT- Open Compute-Optimal Language Models Trained on the Cerebras Wafer-Scale Cluster |
We present qualitative results in Fig. 4, 5,6,7,8,9,10,11,12,13,14,15.
4.2 Qualitative Results
4.3 Ablation Study
Fig. 20 shows a comparison to a model trained without using ControlNet. That model is trained
with exactly same method with Stability’s Depth-to-Image model (Adding a channel to the SD and
continue the training).
Fig. 21 shows the training process. We would like to point out a “sudden convergence phenomenon”
where the model suddenly be able to follow the input conditions. This can happen during the training
process from 5000 to 10000 steps when using 1e-5 as the learning rate.
Fig. 22 shows Canny-edge-based ControlNets trained with different dataset scales.
4.4 Comparison to previous methods
Fig. 14 shows the comparison to Stability’s Depth-to-Image model.
Fig. 17 shows a comparison to PITI [59].
Fig. 18 shows a comparison to sketch-guided diffusion [58].
Fig. 19 shows a comparison to Taming transformer [11].
9
4.5 Comparison of pre-trained models | Adding Conditional Control to Text-to-Image Diffusion Models |
C.22 Enron Emails
To extract the data, we used the mailparser
package25 to extract the body of each email as a
document.
D General Data Processing
This section discusses any processes applied across
multiple datasets.
To combine the constituent datasets, we iterate
until the size of the output dataset is the desired
size, drawing documents from datasets at random,
weighted by the number of documents in each
dataset times the number of epochs desired on
that dataset. Because the number of documents
involved is high, by the law of large numbers, the
number of copies of each dataset present in the Pile
is approximately equal to its epoch count.
Shuffling a dataset posed a major problem due to
our limited memory and computational budget. We
follow Hardin (2018), a method descended from
Rao (1961), and interleave our output to produce
30 output piles. | The Pile- An 800GB Dataset of Diverse Text for Language Modeling |
Diversity. 3DBiCar spans a wide range of 3D biped car-
toon characters, containing 1,500 high-quality 3D models.
First, we carefully collect images of 2D full-body biped car-
toon characters with diverse identities, shape, and textural
styles from the Internet, resulting in 15 character species and
4 image styles, as shown in Fig. 3. Then we recruit six pro-
fessional artists to create 3D corresponding character models
according to the collected reference images. The modeling
result is required to be matched with the reference images
as much as possible. The representative image-model pairs
sampled from our dataset are shown in Fig. 2.
Topological-consistency. The key to building a linear
3 | RaBit- Parametric Modeling of 3D Biped Cartoon Characters with a Topological-consistent Dataset |
Last year you heard us talk about PaLM, which led to many improvements across our products. Today, we’re ready
to announce our latest PaLM model in production: PaLM 2.
PaLM 2 builds on our fundamental research and our latest infrastructure. It’s highly capable at a wide range of
tasks and easy to deploy. We are announcing more than 25 products and features powered by PaLM 2 today.
PaLM 2 models deliver excellent foundational capabilities across a wide range of sizes. We’ve affectionately
named them Gecko, Otter, Bison, and Unicorn. Gecko is so lightweight that it can work on mobile devices: fast
enough for great interactive applications on-device, even when offline. PaLM 2 models are stronger in logic and
reasoning thanks to broad training on scientific and mathematical topics. It’s also trained on multilingual text —
spanning more than 100 languages — so it understands and generates nuanced results. | Google I_O 2023_ Making AI more helpful for everyone |
Computational Resources
Requires computational resources to support
retrieval strategies and technologies related
to databases. External data source integration
and updates need to be maintained.
Preparation and curation of high-quality
training datasets, definition of fine-tuning
objectives, and provision of corresponding
computational resources are necessary.
Latency Requirements
Involves data retrieval, potentially leading to
higher latency.
LLM after fine-tuning can respond without
retrieval, resulting in lower latency.
Reducing Hallucinations
Ethical and Privacy Issues
Inherently less prone to hallucinations as
each answer is grounded in retrieved evi-
dence.
Ethical and privacy concerns arise from
storing and retrieving text from external
databases.
Can help reduce hallucinations by training
the model based on specific domain data but
may still exhibit hallucinations when faced
with unfamiliar input. | Retrieval-AugmentedGenerationforLargeLanguageModels-ASurvey |
We relied on web traffic vs. app traffic to “qualify” companies for the list, as most consumer
GenAI products have been website-first so far (more on this below!). For companies that made
the list that do have a mobile app, we added that traffic, gathered from Sensor Tower as of June
2023, to determine their spot number. Thus, this ranking serves as a tool to identify and
understand category trends, and not as an exhaustive ranking of all consumer AI
platforms.
Here are our top 6 takeaways.
https://a16z.com/how-are-consumers-using-generative-ai/
1/13
19/09/2023, 13:27
How Are Consumers Using Generative AI? | Andreessen Horowitz
TA B L E O F C O N T E N T S | How Are Consumers Using Generative AI_ _ Andreessen Horowitz |
The second line of each test case contains n integers
a_1 , a_2 , ... , a_n (1 <= a_i <= 10^6) .
The second line of each test case contains n integers
a_1 , a_2 , ... , a_n (1 <= a_i <= 10^6) .
It is guaranteed that the sum of n over all test
cases doesn ’t exceed 3 . 10^5.
It is guaranteed that the sum of n over all test
cases doesn ’t exceed 3 . 10^5.
For each test case , print a single integer -- the
maximum possible value of the product from the
statement .
For each test case , print a single integer -- the
minimum possible value of the product from the
statement .
4
3
2 4 3
4
3 2 3 1
2
69 69
6
719313 273225 402638 473783 804745 323328
4
3
2 4 3
4
3 2 3 1
2
69 69
6
719313 273225 402638 473783 804745 323328
Output
Example
Input
Output
12
6
4761
381274500335
Note
Output
Example
Input
Output
8
3
4761
88341292800
Note
Let f(l) =a_l . a_{l+1}
In the first test case ,
Let f(l) =a_l . a_{l+1}
In the first test case ,
* f(1) =
a_1 . a_2
* f(2) = a_2 . a_3 | alphacode |
existing democracies has in fact lived up to the various ideals we might have for
journalism and for democracy. Yet with its many imperfections, at least in
North America and Western Europe, empirical research suggests that
independent, professionally produced news has helped inform the public,
helped people make sense of the world through analysis, interpretation, and
the portrayal of contending forces, and helped members of the public connect
with one another to see themselves as part of a community and act in concert to
influence public affairs (van Zoonen 1998; Curran et al. 2009; Couldry et al.
2010). Beyond these, perhaps the most visible democratic role of news,
investigative journalism specifically can also produce a range of “positive
externalities” that benefit the whole public – even those who do not actually
engage with a particular story – by ensuring more efficient local government,
reducing corruption, and increasing how responsive elected officials are to their | Social_Media_and_Democracy |
and better control the generated music. Note that music in
SymMV is of high quality and can also be directly used for
unconditional music generation without video modality. | VideoBackgroundMusicGeneration |
Fabio Petroni, Aleksandra Piktus, Angela Fan, Patrick
Lewis, Majid Yazdani, Nicola De Cao, James
Thorne, Yacine Jernite, Vladimir Karpukhin, Jean
Maillard, Vassilis Plachouras, Tim Rocktäschel, and
Sebastian Riedel. 2021. KILT: a benchmark for
knowledge intensive language tasks. In Proceedings
of the 2021 Conference of the North American Chap-
ter of the Association for Computational Linguistics:
Human Language Technologies, pages 2523–2544,
Online. Association for Computational Linguistics.
Fabio Petroni, Tim Rocktäschel, Sebastian Riedel,
Patrick Lewis, Anton Bakhtin, Yuxiang Wu, and
Alexander Miller. 2019. Language models as knowl-
In Proceedings of the 2019 Confer-
edge bases?
ence on Empirical Methods in Natural Language
Processing and the 9th International Joint Confer-
ence on Natural Language Processing (EMNLP-
IJCNLP), pages 2463–2473, Hong Kong, China. As-
sociation for Computational Linguistics. | Toolformer |
Limitations. For gender-related errors in translation systems, evaluations do not consider differential harms to people
related to expressing non-binary gender identities (Keyes, 2018; Dev et al., 2021a), or consider contested perspectives
on pronouns across languages and cultures (Lee, 2019). Moreover, while gender agreement into English is amenable to
automatic evaluation, evaluation of gender agreement out of English remains challenging and time-intensive. Finally,
we note that our evaluations focus on only a subset of potential risks (Weidinger et al., 2021), and that our evaluations
focus on model outputs without considering the wider sociotechnical context in which instruction-finetuned language
models exist (Shelby et al., 2023). See Appendix E.8 for measurement quality rubric for this evaluation when translating
into English.
75 | PaLM 2 Technical Report |
planks and 2 sticks on crafting table
return "wooden_axe"
User: [Description] I succeed in step 1, 2, 3, 4, 5.
I finish all steps and I obtain 1 wooden_axe successfully.
==========
User: My current inventory has <inventory>. <visual observation>. How to obtain 1 stone_sword in Minecraft step-by-
step?
Assistant:
Prompt 1: Planning prompt in JARVIS-1
System: Extract the action name, action type, goal object, tool and action rank from the input text.
User: mine({"log":3}, null); # step 1: chop down trees to mine logs
Assistant:
name: mine_log
text condition: chop down trees to mine logs
action: mine
object_item: log
object_number: 3
tool: null
rank: 1
###
input: craft({"planks":12}, {"log":3}, null); # step 2: craft 12 planks from 3 log
Assistant:
name: craft_planks
condition: craft 12 planks from 3 log
16
JARVIS-1: Open-World Multi-task Agents with Memory-Augmented Multimodal Language Models | JARVIS-1 |
B DPO Implementation Details and Hyperparameters
DPO is relatively straightforward to implement; PyTorch code for the DPO loss is provided below:
19
import torch.nn.functional as F
def dpo_loss(pi_logps, ref_logps, yw_idxs, yl_idxs, beta):
"""
pi_logps: policy logprobs, shape (B,)
ref_logps: reference model logprobs, shape (B,)
yw_idxs: preferred completion indices in [0, B-1], shape (T,)
yl_idxs: dispreferred completion indices in [0, B-1], shape (T,)
beta: temperature controlling strength of KL penalty
Each pair of (yw_idxs[i], yl_idxs[i]) represents the
indices of a single preference pair.
"""
pi_yw_logps, pi_yl_logps =
pi_logps[yl_idxs]
ref_yw_logps, ref_yl_logps = ref_logps[yw_idxs], ref_logps[yl_idxs]
pi_logps[yw_idxs],
pi_logratios = pi_yw_logps - pi_yl_logps
ref_logratios = ref_yw_logps - ref_yl_logps
losses = -F.logsigmoid(beta * (pi_logratios - ref_logratios))
rewards = beta * (pi_logps - ref_logps).detach()
return losses, rewards | Direct Preference Optimization |
driving scenario, an LLM selectively activates the required
neural modules by invoking specific functions from the tool
library, ensuring the collection of necessary environmental
information with less redundancy. Upon gathering the
necessary environmental information, the LLM leverages
this data as a query to search in a cognitive memory
for pertinent traffic regulations and the most similar past
driving experience. Finally, the retrieved traffic rules and
driving experience, together with the formerly collected
environmental information, are utilized as inputs to an LLM-
based reasoning engine. The reasoning engine performs
multi-round reasoning based on the inputs and eventually
devises a safe and comfortable trajectory for driving. Our
Agent-Driver architecture harnesses dynamic perception and
prediction capability brought by the tool library, human
knowledge from the cognitive memory, and the strong
decision-making ability of the reasoning engine. This | ALanguageAgentforAutonomousDriving |
26
Same prompt:“room”+ default “a detailed high-quality professional image”Same CFG scale (9.0)Canny EdgeHEDLine (M-LSD)Depth (midas)Normal (from midas)Scribbles (synthesized)Source ImageFigure 24: (Continued) Comparison of six detection types and the corresponding results. The scribble map is extracted from the HED map
with morphological transforms.
27
Same prompt:“robotics”+ default “a detailed high-quality professional image”Same CFG scale (9.0)Canny EdgeHEDLine (M-LSD)Depth (midas)Normal (from midas)Scribbles (synthesized)Source ImageFigure 25: (Continued) Comparison of six detection types and the corresponding results. The scribble map is extracted from the HED map
with morphological transforms.
28 | Adding Conditional Control to Text-to-Image Diffusion Models |
we combine them for the AHWC2S model with input:
= [¯ai,1, . . . , ¯ai,A, hi, wi, cci, cwi , chi].
xAHWC2S
(7)
i
In practice, depending on which measurements are avail-
able, we train and use different regressors. Following the
naming convention of AHWC2S, these models are: AH2S,
AHW2S, AC2S, and AHC2S, as well as their equivalents
without attribute input H2S, HW2S, C2S, and HC2S. For an
evaluation of the contribution of linguistic shape attributes
on top of each anthropometric measurement, see Sup. Mat.
Training Data: To train the A2S and S2A mappings we
use CAESAR data, for which we have SMPL-X shape pa-
rameters, anthropometric measurements, and linguistic at-
tribute scores. We train separate gender-specific models. | Accurate 3D Body Shape Regression using Metric and Semantic Attributes |
This paper provides an investigation of antecedents and consequences of AI’s placebo effect
in HCI. In an experimental study (𝑁 = 65), we examined the influence of negative and positive
verbal AI descriptions. We analyzed the impact of expectations on decision-making in a letter
discrimination task, with or without a sham-AI system.
There are three major shortcomings in the placebo literature in HCI for AI. First, direct effects
on a behavioral level are yet to be found [40, 78]. Second, it is unclear whether nocebo effects,
low expectations impairing behavior, are equally influential as positive expectations based on
verbal descriptions in HCI. Third, we lack a behavioral marker for effectively designing adaptive
AI interfaces that enhance decision-making amidst placebo responses.
Unpublishedworkingdraft.
Notfordistribution. | AI enhance sour performance |
Memory-based Architectures Our document index can be seen as a large external memory for
neural networks to attend to, analogous to memory networks [64, 55]. Concurrent work [14] learns
to retrieve a trained embedding for each entity in the input, rather than to retrieve raw text as in our
work. Other work improves the ability of dialog models to generate factual text by attending over
fact embeddings [15, 13]. A key feature of our memory is that it is comprised of raw text rather
distributed representations, which makes the memory both (i) human-readable, lending a form of
interpretability to our model, and (ii) human-writable, enabling us to dynamically update the model’s
memory by editing the document index. This approach has also been used in knowledge-intensive
dialog, where generators have been conditioned on retrieved text directly, albeit obtained via TF-IDF
rather than end-to-end learnt retrieval [9]. | Retrieval-AugmentedGenerationfor Knowledge-IntensiveNLPTasks |
• Section 4 LLM pre-training: This section explores the various pre-training tech-
niques for LLMs, highlighting how they contribute to resource efficiency. Key
areas such as memory efficiency, data efficiency, and innovative training pipeline
designs are examined, illustrating how each technique impacts the overall resource
utilization during the pre-training phase.
• Section 5 LLM fine-tuning: This section covers the fine-tuning phase of LLMs,
focusing on methods that enhance resource efficiency. It includes detailed discus-
sions on parameter-efficient fine-tuning, which minimizes parameter updates; and
full-parameter fine-tuning, which optimizes the entire parameter set. | Beyond Efficiency |
CNN (32,36)
LSTM (40) DNN (41)
NLP (37,44)
CNN (46) DNN (47)
NB (19) DT (9,20)
DNN [33]
GNN (42) LSTM (40,43)
NLP (38,45)
RL (48) Clustering (23)
DT (50) RNN (28)
GNN (52) FM (53)
ResNet (54)
XGBoost (ensemble) [55]
CNN, RNN [22]
GNN (42) DT (57)
NN (56)
GRU [18]
NLP (24,45)
DT (37)
DRL [30]
DNN (47)
LSTM [29]
CNN (29,36,39)
DNN (35,58) | Knowledge-graph-based explainable AI- A systematic review |
speech of real-world internet users for unjustified removal.
51 Article 1. 2018. Joint Letter on European Commission regulation on online terrorist content.
www.article19.org/resources/joint-letter-on-european-commission-regulation-on-online-ter-
rorist-content/; Reda (2017).
https://doi.org/10.1017/9781108890960 Published online by Cambridge University Press
240
Daphne Keller & Paddy Leerssen
Under a broader view of content moderation, platforms also shape discourse
through the design of their ranking and recommender algorithms, such as
Facebook’s News Feed and YouTube’s Recommended videos (Keller 2019b).
A growing body of literature in computer science and communications science
seeks to ascertain the operation and effects of these complex systems.52 The
design of these algorithms is currently unregulated, but several governments
have recently proposed to do so.53 Most of these initiatives also explicitly
demand greater transparency in algorithmic recommendations.54 | Social_Media_and_Democracy |
large webtext corpora: A case study on the colossal clean crawled corpus, 2021.
Du, N., Huang, Y., Dai, A. M., Tong, S., Lepikhin, D., Xu, Y., Krikun, M., Zhou, Y., Yu, A. W., Firat, O., Zoph, B.,
Fedus, L., Bosma, M., Zhou, Z., Wang, T., Wang, Y. E., Webster, K., Pellat, M., Robinson, K., Meier-Hellstern,
K., Duke, T., Dixon, L., Zhang, K., Le, Q. V., Wu, Y., Chen, Z., and Cui, C. GLaM: Efficient Scaling of Language
Models with Mixture-of-Experts. ICML, 2022. URL https://arxiv.org/abs/2112.06905. | PaLM 2 Technical Report |
data. Organize your data based on your research questions and hypothesis.
4. Display your data based on relationships among the collected data and look for supporting evidence.
5. Cross check your data few times for reliability and validity.
6. So, what did you find from your experimentation? Report without adding any comments of your own.
7. What were the differences? If you are making a comparison. Use T-Test to compare.
8. Analyze your findings to see if it answers your research questions and finds a solution to your problem
statement. Again, avoid making any comments of your own.
Save your energy for the conclusion and discussion chapter. Do not forget to report your results in the
present form because it sounds soothing and original. Example: The interviews indicate that…….result
shows that..
How to Write Your Conclusion and Discussion | How to Write Your PhD Proposal- A Step-By-Step Guide |
Identity management Discrepancy and misalignment between resources of different knowledge graphs is a persistent issue
in current KBX-systems. Managing identities is a prerogative for knowledge-based explainable systems to efficiently use
the available information and avoid undesirable, wide-ranging effects. While a number of principles exist for publishing
and linking resources, a common agreement on what constitutes identical entities is still an open challenge. This also
affect the wide-spread adoption of knowledge graphs in eXplainable AI, that cannot tolerated uncertainty over data quality.
Solutions to this problems, partly investigated [96], could be services to help data modellers and applications to identify
same entities in the real world; better guidelines to correctly use the different types of identity links (e.g. owl:sameAs, | Knowledge graphs as tools for explainable machine learning: A survey |
35
It is not easy to realize the above ideas, though. We’ve mentioned the safety issues of accessing physical
tools, and this is also one main challenge for scientific tool learning since many scientific problems need to
be verified in actual situations, and this process may bring danger if decided by AIs. Meanwhile, foundation
models are generally trained with natural language corpus or natural images, while scientific data are usually
more heterogeneous, numerical, and structured. It is worth exploring how to fuse the general intelligence
learned from plain text and the expertise needed for scientific discovery. Recently, Boiko et al. (2023) show
the potential of this direction and build a system that uses foundation models to design, plan, and execute
scientific experiments (e.g., catalyzed cross-coupling reactions).
6 Conclusion | Tool Learning with Foundation Models |
Gemini models are also capable of operating across modalities and a diverse set of global languages
simultaneously, both for image understanding tasks (e.g., images containing text in Icelandic) and for
generation tasks (e.g., generating image descriptions for a wide range of languages). We evaluate the
performance of generating image descriptions on a selected subset of languages in the Crossmodal-
3600 (XM-3600) benchmark in a 4-shot setting, using the Flamingo evaluation protocol (Alayrac
et al., 2022), without any fine-tuning for all models. As shown in Table 9, Gemini models achieve a
significant improvement over the existing best model, Google PaLI-X.
XM-3600 (CIDER)
English
French
Hindi
Modern Hebrew
Romanian
Thai
Chinese
Average (of 7)
Gemini Ultra
4-shot
86.4
77.9
31.1
54.5
39.0
86.7
33.3
58.4
Gemini Pro
4-shot
87.1
76.7
29.8
52.6
37.7
77.0
30.2
55.9
Google PaLI-X
4-shot
77.8
62.5
22.2
38.7
30.2
56.0
27.7
45.0 | gemini_1_report |
59.0
59.5
5.4 Results on MTEB benchmark
In Table 3, E5 models not only substantially outperform existing ones with similar sizes, but also
match the results of much larger models. The top-2 models on MTEB leaderboard 7 GTRxxl and
Sentence-T5xxl have 4.8B parameters, while our E5large model is more than 10× smaller with 300M
parameters. We expect that our model will benefit from continual scaling up.
Since the difference between BERT-FTbase and E5base is that BERT-FTbase only has fine-tuning
stage, their performance gap demonstrates the usefulness of contrastive pre-training on our proposed
CCPairs dataset. For most task categories except Clustering, performance improves after supervised
fine-tuning. Consistent with prior works [43, 44], this once again demonstrates the importance of
incorporating human knowledge for learning better text embeddings. It remains an open question
whether state-of-the-art embeddings can be obtained in a purely self-supervised manner. | E5 |
24
The Efficiency Spectrum of Large Language Models: An Algorithmic Survey
Efficient LLM Algorithmic Survey, Nov, 2023, USA.
REFERENCES
[1] [n. d.]. Introducing ChatGPT. https://openai.com/blog/chatgpt
[2] [n. d.]. Introducing Claude 2.1. https://www.anthropic.com/index/claude-2-1
[3] [n. d.]. Introducing PyTorch Fully Sharded Data Parallel (FSDP) API. https://pytorch.org/blog/introducing-pytorch-fully-sharded-data-parallel-api/
[4] [n. d.]. Pathways Language Model (PaLM): Scaling to 540 Billion Parameters for Breakthrough Performance. https://blog.research.google/2022/04/pathways-
language-model-palm-scaling-to.html
[5] [n. d.]. Planning for AGI and beyond. https://openai.com/blog/planning-for-agi-and-beyond
[6] Amey Agrawal, Ashish Panwar, Jayashree Mohan, Nipun Kwatra, Bhargav S Gulavani, and Ramachandran Ramjee. 2023. SARATHI: Efficient LLM Inference
by Piggybacking Decodes with Chunked Prefills. arXiv preprint arXiv:2308.16369 (2023). | TheEfficiencySpectrumofLargeLanguageModels-AnAlgorithmicSurvey |
• The voting ties constitute a notable portion to the selection differences between USC and SC,
especially with 8 candidate responses. Specifically, among all responses with the maximum
votes, SC always selects the one with the smallest index, while USC can pick up alternative
ones based on the response format.
• The match ratio between USC and SC consistently surpasses their own task accuracies,
which shows that the consistency criterion is easier to measure than the answer correctness.
• Shifting from 8 to 16 samples, the USC-SC match ratio reduces, suggesting that USC
behaves as an imperfect approximation of SC. However, the difference in response selection
does not always lead to the performance decrease, as USC sometimes selects the correct
response when SC fails.
5 RELATED WORK | UNIVERSALSELF-CONSISTENCYFORLARGELANGUAGEMODELGENERATION |
Among the 328 prompts we evaluated, Claude 2 gave a response judged more harmful than “I can’t help you
with that" in four cases, according to automated evaluation. On manual inspection, in three of the cases its
response did not seem harmful. However, in the other case, the model was disrupted by the jailbreak attempts
in about half of its sampled responses.
3.5 Helpful, Honest, and Harmless (HHH) Evaluations
Anthropic researchers wrote 438 binary choice questions [2, 3, 9] to evaluate language models and preference
models on their ability to identify HHH responses. The model is presented with two outputs and asked to
select the more HHH output. We see in Figure 6 that each of our Claude models is better than the last at this
task 0-shot, showing general improvements in "understanding" helpfulness, honesty, and harmlessness [8].
6 | ClaudeModels |
∗Equal contribution.
1
Universal Self-Consistency for Large Language Model Generation | UNIVERSALSELF-CONSISTENCYFORLARGELANGUAGEMODELGENERATION |
Modeling in Meeting Recognition.. In Interspeech, Vol. 11. 2877–2880.
[90] Tiffany H Kung, Morgan Cheatham, Arielle Medenilla, Czarina Sillos, Lorie De Leon, Camille Elepaño, Maria Madriaga,
Rimel Aggabao, Giezel Diaz-Candido, James Maningo, et al. 2023. Performance of ChatGPT on USMLE: Potential for
AI-assisted medical education using large language models. PLoS digital health 2, 2 (2023), e0000198.
[91] Tom Kwiatkowski, Jennimaria Palomaki, Olivia Redfield, Michael Collins, Ankur Parikh, Chris Alberti, Danielle
Epstein, Illia Polosukhin, Matthew Kelcey, Jacob Devlin, Kenton Lee, Kristina N. Toutanova, Llion Jones, Ming-Wei
Chang, Andrew Dai, Jakob Uszkoreit, Quoc Le, and Slav Petrov. 2019. Natural Questions: a Benchmark for Question
Answering Research. Transactions of the Association of Computational Linguistics (2019). | ASurveyonEvaluationofLargeLanguageModels |
Privacy Preserving Technologies. Personalized tool learning requires models to learn user preferences
from private user information, which inevitably raises privacy-preserving concerns. On the one hand, previous
work has shown that training data extraction attacks can be applied to recover sensitive personal privacy
from foundation models (Carlini et al., 2021), which is a critical challenge for personalized tool learning.
On the other hand, models with high computational costs must be deployed on cloud servers, which require
uploading private data to the cloud to enable personalized responses. It is crucial to develop secure and
trustworthy mechanisms to access and process user data while protecting user privacy. Addressing these
challenges will help unlock the potential of personalized tool learning, enabling more effective and tailored
tool manipulation to meet individual user needs. To this end, it is worth exploring model-oriented distributed | Tool Learning with Foundation Models |
Appendices
The appendix presents supplementary details that extend
beyond the content of the manuscript, aiming to enhance
comprehension of the M2UGen model. Comprehensive
information is provided concerning the model’s training
dataset and training methodology, encompassing explicit
insights into the utilized training approach and the cor-
responding model hyperparameters. Additionally, a thor-
ough exposition is given regarding the composition of the
evaluation sets employed in our study, accompanied by a
delineation of the evaluation methodology and metrics ap-
plied to assess the performance of our model. To elucidate
the diverse capabilities of our model, illustrative demo ex-
amples are also included.
A Music Oriented Dataset Information
We generate 4 different datasets to train the M2UGen
model: MUCaps, MUImage, MUVideo and MUEdit
datasets. The statistics of the datasets are given in Table
7. An example of each from the 4 datasets are shown in
Figure 4. | M2UGen |
not able to "cheat" the mechanism for their own benefit. Moreover, these mechanisms should perform
their computations reasonably (and provably) fast. How to design the trading mechanism in such a
way that these requirements are satisfied? | informatics-phd-projects-2022-23 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.