summaryrefslogtreecommitdiff
path: root/day1/Session-4.tex
diff options
context:
space:
mode:
authorMadhusudan.C.S2009-10-07 00:18:26 +0530
committerMadhusudan.C.S2009-10-07 00:18:26 +0530
commit553da247c42c9d9598264498227dc61cad3fe8e7 (patch)
treec08dfef021c5e22e6c2c6c2b20d205a07c778981 /day1/Session-4.tex
parentc57e38f7958fedd03555c03a1f211dd799c6e34e (diff)
downloadworkshops-more-scipy-553da247c42c9d9598264498227dc61cad3fe8e7.tar.gz
workshops-more-scipy-553da247c42c9d9598264498227dc61cad3fe8e7.tar.bz2
workshops-more-scipy-553da247c42c9d9598264498227dc61cad3fe8e7.zip
Corrected inctime for Session 4 of Day 1.
Diffstat (limited to 'day1/Session-4.tex')
-rwxr-xr-xday1/Session-4.tex10
1 files changed, 6 insertions, 4 deletions
diff --git a/day1/Session-4.tex b/day1/Session-4.tex
index f14bdb8..a0807bd 100755
--- a/day1/Session-4.tex
+++ b/day1/Session-4.tex
@@ -134,6 +134,7 @@
\item \typ{d.keys()} returns a list
\item can we have duplicate keys?
\end{itemize}
+ \inctime{5}
\end{frame}
\begin{frame} {Problem Set 2.1}
@@ -143,7 +144,7 @@
\item[2.1.3] Find the most used Python keywords in your Python code (import keyword).
\end{description}
-\inctime{20}
+\inctime{10}
\end{frame}
\subsection{Set}
@@ -183,6 +184,7 @@ False
>>> len(f10)
5
\end{lstlisting}
+\inctime{5}
\end{frame}
@@ -192,7 +194,7 @@ False
\item[2.2.1] Given a dictionary of the names of students and their marks, identify how many duplicate marks are there? and what are these?
\item[2.2.2] Given a string of the form ``4-7, 9, 12, 15'' find the numbers missing in this list for a given range.
\end{description}
-\inctime{15}
+\inctime{10}
\end{frame}
\subsection{Functions Reloaded!}
@@ -284,7 +286,7 @@ Why is it interesting?\\
\typ{map, reduce, filter}\\
list comprehension\\
generators
- \inctime{10}
+ \inctime{15}
\end{frame}
\subsection{Debugging}
@@ -323,7 +325,7 @@ or modulo by zero
\item Process: Hypothesis, test, refine, rinse-repeat
\item Using \typ{\%debug} and \typ{\%pdb} in IPython
\end{itemize}
- \inctime{10}
+ \inctime{15}
\end{frame}
\begin{frame}[fragile]