Skip to content
Snippets Groups Projects
Commit e4b32b06 authored by Victor's avatar Victor
Browse files

Update for 2025

parent b6da6172
No related branches found
No related tags found
No related merge requests found
Pipeline #21641 failed
Showing
with 357 additions and 159 deletions
...@@ -251,4 +251,4 @@ TSWLatexianTemp* ...@@ -251,4 +251,4 @@ TSWLatexianTemp*
*.sta *.sta
# this thesis # this thesis
thesis.pdf main.pdf
@book{haddadin2013towards,
title={Towards safe robots: approaching Asimov’s 1st law},
author={Haddadin, Sami},
volume={90},
year={2013},
publisher={Springer}
}
@book{goossens94,
author = {Goossens, Michel and Mittelbach, Frank
and Samarin, Alexander},
title = {The LaTeX Companion},
date = {1994},
publisher = {Addison-Wesley},
location = {Reading, Massachusetts},
language = {english},
}
@book{knuth84,
author = {Knuth, Donald E.},
title = {The \TeX book},
date = {1984},
maintitle = {Computers \& Typesetting},
volume = {A},
publisher = {Addison-Wesley},
location = {Reading, Massachusetts},
language = {english},
}
@manual{hanisch14,
author = {Hanisch, Falk},
title = {Ein \LaTeX"=Bundle für Dokumente im neuen Corporate
Design der Technischen Universität Dresden},
date = {2014},
subtitle = {Benutzerhandbuch},
location = {Dresden},
language = {german},
}
\setcounter{chapter}{-1} % sets the number of this chapter to 0
\chapter{Samples}
How to use references to other chapters: In \cref{ch:conclusion} the thesis will be concluded.
In \cref{sec:future-work} the future work will be discussed.
How to use acronym: \ac{CPU} and \ac{GPU}.
\ac{CPU} and \ac{GPU}.
\ac{CPU} and \ac{GPU}.
How to use footnote: The documentation\footnote{tudscr docs: \href{https://ftp.tu-chemnitz.de/pub/tex/macros/latex/contrib/tudscr/doc/tudscr.pdf}{ftp.tu-chemnitz.de/pub/tex/macros/latex/contrib/tudscr/doc/tudscr.pdf}} of the wrapper class might be useful, if there are things you need to understand and are not covered in this short description.
Cite a single work like \cite{walkerVirtualAugmentedMixed2023} or cite multiple works like \cite{suIntegratingVirtualMixed2023,hedayatiImprovingCollocatedRobot2018,kalinovWareVRVirtualReality2021}.
\begin{figure}[h]
\centering
\includegraphics[width=0.8\textwidth]{OMG_MOF_4levels.jpg}
\caption[Bad non-vector image example.]{This is a bad image (pixels visible). Try to use vector images, e.g., using \href{https://excalidraw.com}{excalidraw.com} or \href{https://app.diagrams.net}{app.diagrams.net}.}
\label{fig:mof}
\end{figure}
Also describe every figure in the text, referencing them, e.g., using \cref{fig:mof}.
Math:
$A = \{x | x \in Y\}$
\begin{defs}
A \textbf{Petrinet} is a tuple ${\Sigma = (P, T, F, W)}$.
\label{def:abc}
\end{defs}
Petrinets are defined in \cref{def:abc}.
See at the head of this document how to create your own definitions/lemma environments.
Equation example:
\begin{equation}
x^n + y^n = z^n
\label{eq:example}
\end{equation}
\begin{table}[tb]
\caption{Example of a simple table.}
\label{tab:example}
\centering
\begin{tabular}{llrrr}
\toprule
\textbf{type} &
\textbf{entity} &
\textbf{A} &
\textbf{B} &
\textbf{C} \\
\midrule
T1 & e1 & 27 & 0 & 27 \\
T2 & e2 & 0 & 18 & 88 \\
T2 & e3 & 1 & 27 & 1 \\
\midrule
\multicolumn{2}{r}{Total} & 28 & 45 & 116 \\
\bottomrule
\end{tabular}
\end{table}
\begin{table*}[tb]
\caption{%
Related works comparison with ours.
}
\scriptsize
\centering
\begin{tabularx}{\textwidth}{%
l|c|c|c|
*{3}{c}|
*{2}{c}|
l
}
\toprule
& & & & \multicolumn{3}{c|}{Feat D} & \multicolumn{2}{c|}{Feat E} \\
Works & \rotatebox{90}{Feat A} & \rotatebox{90}{Feat B} & \rotatebox{90}{Feat C} & \rotatebox{90}{Feat D-1} & \rotatebox{90}{Feat D-2} & \rotatebox{90}{Feat D-3} & \rotatebox{90}{Feat E-1} & \rotatebox{90}{Feat E-2} & Notes \\
\midrule
Hedayati et al. \cite{hedayatiImprovingCollocatedRobot2018} & \fullcirc & \emptcirc & \emptcirc & \emptcirc & \fullcirc & \emptcirc & \fullcirc & \emptcirc & \\
Kalinov et al. \cite{kalinovWareVRVirtualReality2021} & \fullcirc & \emptcirc & \emptcirc & \emptcirc & \emptcirc & \fullcirc & \emptcirc & \fullcirc & Some note \\
Su et al. \cite{suIntegratingVirtualMixed2023} & \emptcirc & \emptcirc & \fullcirc & \fullcirc & \emptcirc & \emptcirc & \emptcirc & \fullcirc & \\
Walker et al. \cite{walkerVirtualAugmentedMixed2023} & \emptcirc & \emptcirc & \emptcirc & \fullcirc & \emptcirc & \emptcirc & \emptcirc & \fullcirc & Some other note \\
Ours & \fullcirc & \fullcirc & \fullcirc & \emptcirc & \emptcirc & \fullcirc & \emptcirc & \fullcirc & \\
\bottomrule
\end{tabularx}
\label{tab:related-works}
\end{table*}
The \cref{tab:example} shows a simple table using the booktabs syntax.
Remember to never use vertical lines and nor use double lines.
As a start, \href{https://www.tablesgenerator.com/}{tablesgenerator.com} or other tools can be used.
\begin{figure}[t!]
\begin{lstlisting}[language=TeX]
\begin{table}[tb]
\caption{Example of a simple table.}
\centering
\begin{tabular}{llrrr}
\toprule
\textbf{type} &
\textbf{entity} &
\textbf{A} &
\textbf{B} &
\textbf{C} \\
\midrule
T1 & e1 & 27 & 0 & 27 \\
T2 & e2 & 0 & 18 & 88 \\
T2 & e3 & 1 & 27 & 1 \\
\midrule
\multicolumn{2}{r}{Total} & 28 & 45 & 116 \\
\bottomrule
\end{tabular}
\label{tab:example}
\end{table}
\end{lstlisting}
\caption{Code for the example \cref{tab:example}.}
\label{fig:code_for_tab}
\end{figure}
\chapter{Introduction}
\label{ch:introduction}
The chapter should broadly contextualize your research and motivate your work.
\section{Motivation}
Why you are doing your research.
\section{Goal}
Clarify what should be achieved by the work.
\section{Structure of the Thesis}
Here the structure can be \emph{briefly} explained.
\chapter{Background}
\label{ch:background}
\chapter{Related Work}
\label{ch:related-work}
Literature review chapter.
Describing the research field with relevant works on the same or similar issues.
\chapter{Methods and Implementation}
\label{ch:methods-and-implementation}
\chapter{Evaluation}
\label{ch:evaluation}
\begin{itemize}
\item Describe objectives of evaluation
\item Present structure and used material of evaluation
\item Show and explain results of evaluation
\end{itemize}
\chapter{Conclusion} \chapter{Conclusion}
\label{ch:conclusion} \label{ch:conclusion}
\blindtext \section{Summary}
\section{Future Work}
\label{sec:future-work}
\begin{abstract}[pagestyle=empty.tudheadings]
This is an abstract. The abstract is written after finishing the work and should give an overview about the motivation, used methods, as well as the results. It is here to inform the reader about the core topics of the work and if it is relevant to his research. The abstract stands for itself and uses no components of the rest of the work. In consequence, there are no references nor citations used here. There should always be an english version of your abstract, regardless of the language the work is actually written in.
\end{abstract}
\chapter{Symbols and Acronyms}
\label{ch:acronyms}
\begin{multicols}{2}
\begin{acronym}[CPU] % put the longest acronym here
\acro{CPU}{Central Processing Unit}
\acro{GPU}{Graphics Processing Unit}
\end{acronym}
\end{multicols}
\chapter{Appendix} \chapter{Appendix I}
\label{sec:appendix-1}
Here, you can put long listings or informative figures not important to understand the main thesis. Here, you can put long listings or informative figures not important to understand the main thesis.
\chapter{Appendix II}
\label{sec:appendix-2}
\begin{declarations}[pagestyle=empty.tudheadings]
\confirmation
\end{declarations}
syn Collection<Sensor> Route.requiredSensors() {
java.util.List<Sensor> sensors = new ArrayList<>();
for (Ref sensorRef : getRequiredSensorList()) {
sensors.add(sensorRef.asSensor());
}
return sensors;
}
File moved
\usepackage[T1]{fontenc} \usepackage[T1]{fontenc}
\usepackage{hyphsubst}
\usepackage{gentium}
\usepackage{inconsolata}
\usepackage{babel} \usepackage{babel}
\usepackage{isodate} \usepackage{isodate}
\setkomafont{disposition}{\normalcolor\bfseries\sffamily} \setkomafont{disposition}{\normalcolor\bfseries\sffamily}
\usepackage[ \usepackage[
style=numeric-comp,
backend=biber, backend=biber,
url=false, style=numeric-comp,
doi=false,
isbn=false, isbn=false,
hyperref, hyperref,
]{biblatex} ]{biblatex}
\addbibresource{bibliography.bib} \addbibresource{references.bib}
\AtEveryBibitem{%
\clearfield{note}%
}
\usepackage{hyperref} % makes all links clickable \PassOptionsToPackage{hyphens}{url}
\usepackage[hidelinks]{hyperref} % makes all links clickable
\usepackage[obeyFinal,textwidth=30mm,colorinlistoftodos,prependcaption,textsize=tiny]{todonotes}
\usepackage{graphicx} \usepackage{graphicx}
\usepackage{blindtext} % add blindtext \graphicspath{{figs/}{./}}
\usepackage[babel]{microtype} % Improve font layout (spacing, hyphenation etc.) \usepackage[babel]{microtype} % improve font layout (spacing, hyphenation etc.)
\usepackage{doi} % Add links for DOIs, handle "_" in dois \usepackage{doi} % add links for DOIs, handle "_" in dois
\usepackage{csquotes} % provide \enquote \usepackage{csquotes} % provide \enquote
\usepackage{fnpct} % punctuation, especially together with footnotes \usepackage{fnpct} % punctuation, especially together with footnotes
\usepackage{ifdraft} % for detecting draft/final mode
\usepackage{version} % conditionally to exclude parts of document
\usepackage{tcolorbox} % make colored boxes \usepackage{tcolorbox} % make colored boxes
\usepackage{caption} \usepackage{caption}
\usepackage{subcaption} \usepackage{subcaption}
\usepackage{booktabs} % must-have for tables
\usepackage{tabularx} % tabularx environment \usepackage{tabularx} % tabularx environment
\usepackage{booktabs} % must-have for tables
\usepackage{multirow} % for combining multiple rows \usepackage{multirow} % for combining multiple rows
\usepackage{makecell} % enables line breaks in cells \usepackage{makecell} % enables line breaks in cells
\usepackage{quoting} \usepackage{acronym}
\usepackage{multicol} % use multiple columns, used for the acronyms section
\usepackage{xfrac} \usepackage[capitalize]{cleveref}
\input{lst.tex}
\usepackage[capitalise,nameinlink,noabbrev]{cleveref} % automatically inserts Fig. X in the text with \cref{..}
% if you need mathy stuff % cleveref settings, if you need mathy stuff
\newtheorem{lem}{Lemma} \newtheorem{lem}{Lemma}
\crefname{lem}{Lemma}{Lemmas} \crefname{lem}{Lemma}{Lemmas}
\newtheorem{thm}{Theorem} \newtheorem{thm}{Theorem}
...@@ -58,13 +43,24 @@ ...@@ -58,13 +43,24 @@
\newtheorem{defs}{Definition} \newtheorem{defs}{Definition}
\crefname{defs}{Def.}{Defs.} \crefname{defs}{Def.}{Defs.}
\newenvironment{draftcontent}[1][,]{% \usepackage{listings} % for displaying code
\begin{tcolorbox}[title=Draft of content,#1]\begin{itemize}}{% \definecolor{KeywordBlue}{cmyk}{0.88,0.77,0,0}
\end{itemize}\end{tcolorbox}} \definecolor{CommentGreen}{cmyk}{0.87,0.24,1.0,0.13}
\lstset{
basicstyle=\scriptsize\ttfamily,
tabsize=2,
language=C,
commentstyle=\color{CommentGreen},
keywordstyle=\ttfamily\color{KeywordBlue},
backgroundcolor=\color[rgb]{0.95,0.95,0.95},
breaklines=true,
literate={\\\%}{{\textcolor{black}{\\\%}}}1}
% Exclude draftcontent environment, if "final" is selected % For comparison table
\ifoptionfinal{ \newcommand*\emptcirc[1][1ex]{\tikz[baseline={(0,-0.1)}]\draw (0,0) circle (#1);}
\excludeversion{draftcontent} \newcommand*\halfcirc[1][1ex]{%
}{ \begin{tikzpicture}[baseline={(0,-0.1)}]
\includeversion{draftcontent} \draw[fill] (0,0)-- (90:#1) arc (90:270:#1) -- cycle ;
} \draw (0,0) circle (#1);
\end{tikzpicture}}
\newcommand*\fullcirc[1][1ex]{\tikz[baseline={(0,-0.1)}]\fill (0,0) circle (#1);}
\usepackage{listings}
\usepackage{inconsolata}
\lstdefinestyle{common-style}{
basicstyle=\scriptsize\ttfamily, % the size of the fonts that are used for the code
showspaces=false, % show spaces adding particular underscores
showstringspaces=false, % underline spaces within strings
showtabs=false, % show tabs within strings adding particular underscores
% frame=tlrb, % adds a frame around the code
framexleftmargin=1em, % space between left part of frame and listing
tabsize=2, % sets default tabsize to 2 spaces
breaklines=true, % sets automatic line breaking
breakatwhitespace=true, % sets if automatic breaks should only happen at whitespace
keywordstyle={\color{blue}\textbf}, % keywords are blue
commentstyle={\color{gray}}, % comments
literate={\$}{{{\$}}}1,
basewidth=0.5em,
breakindent=40pt,
breakautoindent=true,
escapechar=\&,
aboveskip={0.1\baselineskip}
}
\lstdefinestyle{shortlisting}{
xleftmargin=\parindent,
frame=none,
aboveskip=3pt,belowskip=3pt
}
\lstdefinestyle{unboxed}{
style=common-style,
frame=none,
}
% JastAdd
\lstdefinelanguage{AST}{
style=common-style,
morekeywords={abstract,rel},
otherkeywords={::=,->,<,>},
morecomment=[l]{//}, morecomment=[s]{/*}{*/},
}
\lstdefinelanguage{JRAG}[]{java}{
style=common-style,
morekeywords={abstract,public,private,boolean,aspect,null,syn,inh,coll,eq,with,int,contributes,new,return,for,if,else,this,to,true,false},
morecomment=[l]{//}, morecomment=[s]{/*}{*/},
}
\newcommand{\lstbg}[3][0pt]{{\fboxsep#1\colorbox{#2}{\strut #3}}}
\lstdefinelanguage{diff}[]{java}{
style=common-style,
morecomment=[f][\lstbg{HKS07!30}]-,
morecomment=[f][\lstbg{HKS65!30}]+,
morecomment=[f][\textit]{@@},
%morecomment=[f][\textit]{---},
%morecomment=[f][\textit]{+++},
}
\lstdefinestyle{AST} { language=AST,style=common-style }
\lstdefinestyle{JRAG} { language=JRAG,style=common-style }
\lstdefinestyle{Java} { language=Java,style=common-style }
main.tex 0 → 100644
\documentclass[
a4paper,
ngerman, english, % the main language as last
cdgeometry=symmetric,
DIV=10,
cdfont=no,
cleardoublespecialpage=true,
open=right,
BCOR=6mm,
oneside, % use 'twoside' to print the document
numbers=noenddot,
]{tudscrbook}
\input{header}
\begin{document}
\faculty{Faculty of Computer Science}
\institute{SMT}
\chair{%
Chair of Software Technology
% Junior Professorship in Immersive Media
}
\title{%
The Title of the Work
}
% \thesis{bachelor}
% \graduation[B.Sc.]{Bachelor of Science}
\thesis{master}
\graduation[M.Sc.]{Master of Science}
% \thesis{diploma}
% \graduation[Dipl.Inf.]{Diplom-Informatiker}
\author{Firstname Lastname}
\matriculationnumber{12345678}
\dateofbirth{01.01.1990}
\referee{%
Prof. Dr. rer. nat. Uwe Aßmann \and
Jun.-Prof. Dr. Matthew McGinity
% Prof. Dr. rer. nat. Stefan Gumhold
% Dr.-Ing. Lars Engeln
}
\supervisor{Victor Victor}
\date{01.01.2025}
\maketitle
\frontmatter
\input{chapters/abstract}
\input{chapters/confirmation}
\tableofcontents
\input{chapters/acronyms}
\mainmatter
\input{chapters/0-samples} % comment or delete this line
\input{chapters/1-introduction}
\input{chapters/2-background}
\input{chapters/3-related-work}
\input{chapters/4-methods-and-implementation}
\input{chapters/5-evaluation}
\input{chapters/6-conclusion}
\printbibliography[heading=bibintoc]
\listoffigures
\listoftables
\appendix
\input{chapters/appendix}
\end{document}
@article{suIntegratingVirtualMixed2023,
title = {Integrating {{Virtual}}, {{Mixed}}, and {{Augmented Reality}} into {{Remote Robotic Applications}}: {{A Brief Review}} of {{Extended Reality-Enhanced Robotic Systems}} for {{Intuitive Telemanipulation}} and {{Telemanufacturing Tasks}} in {{Hazardous Conditions}}},
shorttitle = {Integrating {{Virtual}}, {{Mixed}}, and {{Augmented Reality}} into {{Remote Robotic Applications}}},
author = {Su, Yun-Peng and Chen, Xiao-Qi and Zhou, Cong and Pearson, Lui Holder and Pretty, Christopher G. and Chase, J. Geoffrey},
year = {2023},
month = nov,
journal = {Applied Sciences},
volume = {13},
number = {22},
pages = {12129},
issn = {2076-3417},
doi = {10.3390/app132212129},
urldate = {2024-06-05},
abstract = {There is an increasingly urgent need for humans to interactively control robotic systems to perform increasingly precise remote operations, concomitant with the rapid development of space exploration, deep-sea discovery, nuclear rehabilitation and management, and robotic-assisted medical devices. The potential high value of medical telerobotic applications was also evident during the recent coronavirus pandemic and will grow in future. Robotic teleoperation satisfies the demands of the scenarios in which human access carries measurable risk, but human intelligence is required. An effective teleoperation system not only enables intuitive human-robot interaction (HRI) but ensures the robot can also be operated in a way that allows the operator to experience the ``feel'' of the robot working on the remote side, gaining a ``sense of presence''. Extended reality (XR) technology integrates real-world information with computer-generated graphics and has the potential to enhance the effectiveness and performance of HRI by providing depth perception and enabling judgment and decision making while operating the robot in a dynamic environment. This review examines novel approaches to the development and evaluation of an XR-enhanced telerobotic platform for intuitive remote teleoperation applications in dangerous and difficult working conditions. It presents a strong review of XR-enhanced telerobotics for remote robotic applications; a particular focus of the review includes the use of integrated 2D/3D mixed reality with haptic interfaces to perform intuitive remote operations to remove humans from dangerous conditions. This review also covers primary studies proposing Virtual Reality (VR), Augmented Reality (AR), and Mixed Reality (MR) solutions where humans can better control or interact with real robotic platforms using these devices and systems to extend the user's reality and provide a more intuitive interface. The objective of this article is to present recent, relevant, common, and accessible frameworks implemented in research articles published on XR-enhanced telerobotics for industrial applications. Finally, we present and classify the application context of the reviewed articles in two groups: mixed reality--enhanced robotic telemanipulation and mixed reality--enhanced robotic tele-welding. The review thus addresses all elements in the state of the art for these systems and ends with recommended research areas and targets. The application range of these systems and the resulting recommendations is readily extensible to other application areas, such as remote robotic surgery in telemedicine, where surgeons are scarce and need is high, and other potentially high-risk/high-need scenarios.},
copyright = {https://creativecommons.org/licenses/by/4.0/},
langid = {english}
}
@inproceedings{hedayatiImprovingCollocatedRobot2018,
title = {Improving {{Collocated Robot Teleoperation}} with {{Augmented Reality}}},
booktitle = {Proceedings of the 2018 {{ACM}}/{{IEEE International Conference}} on {{Human-Robot Interaction}}},
author = {Hedayati, Hooman and Walker, Michael and Szafir, Daniel},
year = {2018},
month = feb,
pages = {78--86},
publisher = {ACM},
address = {Chicago IL USA},
doi = {10.1145/3171221.3171251},
urldate = {2025-04-13},
isbn = {978-1-4503-4953-6},
langid = {english}
}
@inproceedings{kalinovWareVRVirtualReality2021,
title = {{{WareVR}}: {{Virtual Reality Interface}} for {{Supervision}} of {{Autonomous Robotic System Aimed}} at {{Warehouse Stocktaking}}},
shorttitle = {{{WareVR}}},
booktitle = {2021 {{IEEE International Conference}} on {{Systems}}, {{Man}}, and {{Cybernetics}} ({{SMC}})},
author = {Kalinov, Ivan and Trinitatova, Daria and Tsetserukou, Dzmitry},
year = {2021},
month = oct,
pages = {2139--2145},
issn = {2577-1655},
doi = {10.1109/SMC52423.2021.9659133},
urldate = {2025-04-13},
abstract = {WareVR is a novel human-robot interface based on a virtual reality (VR) application to interact with a heterogeneous robotic system for automated inventory management. We have created an interface to supervise an autonomous robot remotely from a secluded workstation in a warehouse that could benefit during the current pandemic COVID-19 since the stocktaking is a necessary and regular process in warehouses, which involves a group of people. The proposed interface allows regular warehouse workers without experience in robotics to control the heterogeneous robotic system consisting of an unmanned ground vehicle (UGV) and unmanned aerial vehicle (UAV). WareVR provides visualization of the robotic system in a digital twin of the warehouse, which is accompanied by a real-time video stream from the real environment through an onboard UAV camera. Using the WareVR interface, the operator can conduct different levels of stocktaking, monitor the inventory process remotely, and teleoperate the drone for a more detailed inspection. Besides, the developed interface includes remote control of the UAV for intuitive and straightforward human interaction with the autonomous robot for stocktaking. The effectiveness of the VR-based interface was evaluated through the user study in a ``visual inspection'' scenario.},
keywords = {Autonomous aerial vehicles,Inspection,Real-time systems,Robot vision systems,Streaming media,Virtual reality,Visualization}
}
@article{walkerVirtualAugmentedMixed2023,
title = {Virtual, {{Augmented}}, and {{Mixed Reality}} for {{Human-robot Interaction}}: {{A Survey}} and {{Virtual Design Element Taxonomy}}},
shorttitle = {Virtual, {{Augmented}}, and {{Mixed Reality}} for {{Human-robot Interaction}}},
author = {Walker, Michael and Phung, Thao and Chakraborti, Tathagata and Williams, Tom and Szafir, Daniel},
year = {2023},
month = jul,
journal = {J. Hum.-Robot Interact.},
volume = {12},
number = {4},
pages = {43:1--43:39},
doi = {10.1145/3597623},
urldate = {2024-12-17},
abstract = {Virtual, Augmented, and Mixed Reality for Human-Robot Interaction (VAM-HRI) has been gaining considerable attention in HRI research in recent years. However, the HRI community lacks a set of shared terminology and framework for characterizing aspects of mixed reality interfaces, presenting serious problems for future research. Therefore, it is important to have a common set of terms and concepts that can be used to precisely describe and organize the diverse array of work being done within the field. In this article, we present a novel taxonomic framework for different types of VAM-HRI interfaces, composed of four main categories of virtual design elements (VDEs). We present and justify our taxonomy and explain how its elements have been developed over the past 30 years as well as the current directions VAM-HRI is headed in the coming decade.}
}
\chapter{Background}
\label{ch:basics}
\input{sections/figures}
\input{sections/install}
\section{What is ABC?}
\blindtext
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Please register or to comment