Commit 0c3b5cc5 authored by Marc Feger's avatar Marc Feger
Browse files

Refactor slides/dataset.tex and master.tex

parent 98fdbf2e
Pipeline #44630 passed with stage
in 3 minutes and 1 second
......@@ -21,8 +21,7 @@
\usepackage{mathabx}
\usepackage[linesnumbered,algoruled,boxed,lined]{algorithm2e}
\usepackage{amssymb}
\usepackage[square,numbers]{natbib}
\usepackage[longnamesfirst, authoryear]{natbib}
\usepackage{textcomp}
\usepackage{listings}
\usepackage{color}
......@@ -183,7 +182,7 @@ Department for Computer-Networks and Communication-Systems}
% % % % % % % % % Ende der eingefügten LaTeX-Dateien % % % % % % % % %
\begin{frame}[allowframebreaks]\frametitle{References}
\bibliographystyle{abbrvnat}
\bibliographystyle{plainnat}
\bibliography{references}
\nocite{*}
\end{frame}
......
......@@ -4,16 +4,10 @@
\begin{frame}
\frametitle{Corpus}
\begin{itemize}
\item For our study, we used the Webis-ArgRank2017 dataset from Wachsmuth et
al.\footnote[1]{H. Wachsmuth, B. Stein, and Y. Ajjour "PageRank" for Argument
Relevance}
\item In this dataset Wachsmuth et al. constructed a ground-truth argument graph
as well as benchmark for argument ranking
\item The data are originally collected from the Argument Web and stored in an
argument graph
\item The Argument Web was the largest existing argument database with a
structured argument corpora at that time
\item For our study, we used the Webis-ArgRank2017 dataset by \cite{wachsmuth:2017a}.
\item Webis-ArgRank2017 contains a ground-truth argument graph
as well as a benchmark for argument ranking
\item The data was originally collected within the Argument-Web
\end{itemize}
\end{frame}
......@@ -22,17 +16,16 @@
\begin{itemize}
\item In the resulting argument graph G = (A, E)
\begin{itemize}
\item Each node represents $a_i \in A$ an argument consisting of a conclusion
\item Each node represents an argument $a_i \in A$ consisting of a conclusion
$c_i$ and a not-empty set of premises $P_i$ $\Rightarrow$ $a_i = \langle c_i, P_i \rangle$
\item An edge $(a_j, a_i)$ is given if the conclusion $a_j$ is used as a premise
of $a_i$
\item Consequently, $P_i = \{c_1,...,c_k\}, k \geq 1$
\end{itemize}
\end{itemize}
\begin{figure}
\includegraphics[width=0.4\linewidth]{bilder/DatasetLocalView2.png}
\caption{Argument Graph from Argument Web}
\includegraphics[width=0.7\linewidth]{bilder/DatasetLocalView2.png}
\caption{Part of the Argument-Web used to create Webis-ArgRank2017.}
\end{figure}
\end{frame}
......@@ -40,20 +33,16 @@
\subsection{Benchmark Argument Ranking}
\begin{frame}
\frametitle{Benchmark Argument Ranking}
\cite{wachsmuth:2017a} kept those arguments fulfilling:
\begin{itemize}
\item To create the benchmark dataset, Wachsmuth et al. only kept arguments
from the graph that fulfill their requirements
\begin{itemize}
\item A conclusion must be part in more than one argument
\item All nodes are removed that do not contain a
real claim
\item Additionally, an argument:
\begin{itemize}
\item has to be a valid-counter argument
\item must be based on reasonable premises
\item must allow a logic interference to be drawn
\end{itemize}
\end{itemize}
\item The conclusion was part in more than one argument
\item Real claim
\end{itemize}
Additionally, an argument had to be:
\begin{itemize}
\item a valid counterargument
\item based on reasonable premises
\item logically consistent
\end{itemize}
\end{frame}
......@@ -74,15 +63,7 @@
\begin{frame}
\frametitle{Evaluation Method}
\begin{itemize}
\item Wachsmuth et al. used Kendall's $\tau$ to evaluate the agreement between
the experts and to ensure comparability
\item Kendall $\tau$ is a correlation coefficient that indicates the agreement
between two quantities with respect to a property
\begin{itemize}
\item In this case, this means the agreement between two experts with respect to
an argument
\item -1 signifies a complete disagreement and +1 a complete agreement
\end{itemize}
\item The mean over all experts for the evaluation of the benchmark is 0.36
\item \cite{wachsmuth:2017a} used Kendall's $\tau \in [-1, 1]$ to evaluate the agreement between the experts and to ensure comparability
\item The mean over all experts for the evaluation of the benchmark is $\tau \approx 0.36$
\end{itemize}
\end{frame}
\ No newline at end of file
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment