\\ Corning Community College \\ UNIX/Linux Fundamentals \\ \\ Lab 0xA: Data Analysis with Regular Expressions and Scripting \\ \\ ~~TOC~~ =====Objective===== To continue to build on our knowledge of Regular Expressions, Scripting, and utilizing them to solve problems. =====Reading===== Referencing the following manual pages will likely be of considerable value: * **grep**(**1**) * **sed**(**1**) * **awk**(**1**) =====Note===== This lab is involved. Information obtained in early steps are built upon with increasingly complex functionality. If at any point something doesn't make sense, or you aren't getting the output you think you should be getting- **ask**. It is your responsibility to understand what is going on, please be proactive by asking questions (mailing list, IRC, etc.). =====Background===== As we've been exploring Regular Expressions, Shell Scripting, and even the various tools on the system, you've been told that these are important building blocks to aid you in more effective problem solving. Now, we've amassed a considerable amount of knowledge, we can begin to see the fruits of our labor, and will put our skills to use solving a problem that might previously have been considered "difficult" or near "impossible". =====Problem Description===== As students at CCC, a routine activity that takes place each semester is the selection of classes for the following semester. To put a schedule together, courses must be looked up and a selection of compatible times are selected, sometimes choosing from a selection of offerings, and ultimately a CRN (Course Reference Number) must be identified in order to communicate to the system. On the main [[http://www.corning-cc.edu/|CCC web site]] is a link entitled [[https://ssbprod.corning-cc.edu/PROD/bwckschd.p_disp_dyn_sched|Class Schedule]], which allows us to select a semester and perform searches for available courses. As it turns out, this functionality generates data in HTML format, yet it contains all the useful information we might possibly need for course selections. This is one of those perfect examples that can be solved with our UNIX skills... the data we find available to us is in a form not immediately readable for our needs.. so what do we do when the universe doesn't align to our needs? We realign the universe! =====Obtain the Data===== In preparation for this exercise, I have taken the liberty of downloading a class listing for the Fall 2013 semester. This list contains all the courses offered at the primary college locations and Internet courses, and excludes ACE courses, and courses taught at non-primary college locations (high schools, etc.). This file can be found in the **courselist/** subdirectory of the UNIX Public Directory. It is called: **fall2013-20130417.html.gz** We'll want to copy this file to your home directory. ^ 1. ^|Do the following:| | ^ a.|Copy the indicated file to your home directory. How did you do this?| |:::^ b.|List the file. How large is it?| |:::^ c.|What type of file is it? How did you determine this?| |:::^ d.|We want the HTML data, so unravel this file to obtain that data. How did you do this?| |:::^ e.|How large is the HTML data?| |:::^ f.|What is the compression ratio achieved with this data?| Now that we have a copy of the data, we can move on to studying it. =====Analyzing the Raw Data===== The first step we must take when tackling a problem like this is to get an understanding of the data we are working with. Regular Expressions are cool and all, but they aren't useful unless we know what it is we are describing. Our first task is to locate any common patterns in our data that we might be able to use to our advantage with Regular Expressions. ^ 2. ^|Viewing the HTML file in **vi**, answer me the following:| | ^ a.|This file contains courses offered next semester. Search for the course entry for "CSIT 2044". How did you do this?| |:::^ b.|Comparing the data in this file, is there any similarity to an "ENGL 1020" course? How about a "MATH 1230" course? Is there any pattern in common among all the courses?| As it stands, each course has an information string as follows (I'll use UNIX as an example): ^ UNIX/Linux Fundamentals - 92629 - CSCS 1730 - 001 | After the initial HTML data, we get actually course data we are interested in... there's a pattern we can take advantage of here: the course information is separated into fields, and each field is separated by a hyphen "-". In its default state, the data is arranged as follows: - Course Title - Course Reference Number (CRN) - Course Prefix/Number - Course Section Check out some other courses and verify that this pattern holds true. The actual data will vary, but the pattern/presentation of the data should be identical. Because of that, we can describe it with a Regular Expression pattern, and perform manipulations on it. =====Isolating the Course Information Strings===== Although there's additional information (room, time, instructor, credits), let's start off by isolating all the individual course information strings (that exhibit that pattern above). Using the UNIX class again as an example, the actual line in question is as follows: UNIX/Linux Fundamentals - 92629 - CSCS 1730 - 001 If we go and look at another class, say ARTS 1030, we see the following: Drawing I - 92630 - ARTS 1030 - 001 and GOVT 1010, we see the following: American Federal Government - 93193 - GOVT 1010 - 001 In context, these lines are surrounded by other lines of information, which we aren't immediately interested in. For example, looking at HUMA 1020:

Basic Humanities II - 92249 - HUMA 1020 - 001 Associated Term: Fall 2013
Registration Dates: Mar 25, 2013 to Dec 15, 2013
Each class should be in a similar situation. The line containing the course information is surrounded by lines that contain other information (whether useful or useless, there is other data there than what we are presently interested in locating). ^ 3. ^|Through analyzing the data, answer me the following:| | ^ a.|If we wanted to perform a search that would only hit the course information lines (ie a pattern that would match just that line, and match that line for each course in the file), what does the RegEx pattern look like?| |:::^ b.|Perform the search in **vi** (using **/**, verify that it hits that line in some course). Does it snap to the appropriate line?| |:::^ c.|Hit **n** to go to the next match. And hit **n** again. And again. Are you consistently hitting the course information line for each course?| You absolutely need to have a correctly working pattern in order to proceed. If you have ANY questions, please ask them. This lab will fail to cooperate with you if your pattern is not adequate. Return to the command prompt. Time to start prototyping our solution. We'll want to come up with a command-line that isolates those course information lines for us. A prototype for that command-line will look something like this (substitute your working RegEx pattern in place of the string "REGEX" in the example below): lab46:~$ cat fall2013-20110417.html | grep 'REGEX' | less When you put in the same pattern you came up with while searching in **vi**, your screen should be filled with data that looks like this (and much much more): Accounting Practices - 92473 - ACCT 1000 - 001 Learning Outcomes Accounting Practices - 92467 - ACCT 1000 - 002 Learning Outcomes Accounting Practices - 92938 - ACCT 1000 - 003 Learning Outcomes Accounting Practices - 92789 - ACCT 1000 - 005 Learning Outcomes Accounting Practices - 92792 - ACCT 1000 - 006 Learning Outcomes Accounting Practices - 92980 - ACCT 1000 - 007 Learning Outcomes : Because we piped our output to **less**(**1**), it stops after the first screenful of information. Pressing the down/up arrow keys or the space bar will navigate us through this data. What we're interested in at this point is that the data that is being produced all seems to match those lines in the file that contain the course information string. =====Filtering unnecessary data===== When you're satisfied with the information your pattern and resultant **grep** search has produced, our next step is to refine the information-- to make it more readable. To do this, we will make use of the **sed**(**1**) utility, which is a //steam editor//; it allows us to take the output and perform edits on it, much like we could in a text editor, only we specify on the command-line the actual work we wish to perform. If you recall from our explorations of **vi**, it has a //search and replace// capability that proved to be rather powerful. **sed**(**1**) also possesses this ability, and we can unlock it as follows: cat FILE | grep 'REGEX' | sed 's/PATTERN/REPLACEMENT/g' | less Of specific interest to us are the **s**, **PATTERN**, **REPLACEMENT** and **g** options to **sed**(**1**). They have the following functionality: * **s**: Invoke the **sed**(**1**) //search and replace// command. By default the forward slash **/** is the field separator. * **PATTERN**: The first field following the search command is the pattern we are looking for. In this case, we want to come up with a new pattern that will match a portion of the text we wish to get rid of. * **REPLACEMENT**: This field is what we wish to replace any matched text with. * **g**: Not necessarily needed in all cases, the **g** indicates we wish to perform this search and replace globally for **all** occurrences on a line. I'd recommend getting in the habit of using it, and then recognizing when you don't want to use it. So, looking at the data leading up to the course information we're interested in, can we come up with a pattern to describe it? I think so. ^ 4. ^|Craft a RegEx pattern that does the following:| | ^ a.|Starts at the beginning of the line.| |:::^ b.|Goes until it encounters some unique text just before our desired information.| |:::^ c.|Specifically describe the pattern of the data just before our desired information.| |:::^ d.|What is your pattern?| To test your pattern, you'll want to do the following: lab46:~$ cat fall2013-20110417.html | grep 'REGEX' | sed 's/PATTERN//g' | less Where **PATTERN** is a new Regular Expression pattern that successfully matches the beginning of the lines we're interested in (actually all that **grep** is producing at this point), and replacing it with nothing (the two consecutive slashes indicate we're not interested in replacing the matched data with anything). If successful your output should appear as follows (note this is showing fall 2011 data, so the CRNs and number of offerings may be different): Accounting Practices - 81559 - ACCT 1000 - 001 Accounting Practices - 82350 - ACCT 1000 - 003 Financial Accounting - 82355 - ACCT 1030 - 001 Financial Accounting - 81558 - ACCT 1030 - 002 Financial Accounting - 81107 - ACCT 1030 - 003 Financial Accounting - 81108 - ACCT 1030 - 004 Financial Accounting - 81173 - ACCT 1030 - 005 Financial Accounting - 82115 - ACCT 1030 - 006 Managerial Accounting - 82078 - ACCT 1040 - 003 Accounting Procedures - 81123 - ACCT 1050 - 001 Federal Income Tax - 81783 - ACCT 1100 - 001 Federal Income Tax - 82358 - ACCT 1100 - 002 Intermediate Accounting I - 81124 - ACCT 2030 - 001 Intermediate Accounting I - 82359 - ACCT 2030 - 002 Computerized Accounting - 82361 - ACCT 2100 - 001 Cultural Anthropology - 81139 - ANTH 2120 - 001 Elem Mod Stand Arabic Con&StrI - 82296 - ARAB 1010 - 001 Elem Mod Arabic Con&StrI - lab - 82297 - ARAB 1010 - 071 Introduction Art Appreciation - 81505 - ARTS 1004 - 002 Drawing I - 81771 - ARTS 1030 - 001 Drawing I - 82176 - ARTS 1030 - 002 Drawing I - 82112 - ARTS 1030 - 003 Drawing I - 81503 - ARTS 1030 - 004 Ceramics I - 81151 - ARTS 1210 - 001 Ceramics I - 82110 - ARTS 1210 - 002 Ceramics I - 81504 - ARTS 1210 - 003 Ceramics I - 82134 - ARTS 1210 - 004 Ceramics I - 81176 - ARTS 1210 - 005 Basic Black & White Photo - 81873 - ARTS 1220 - 001 Basic Black & White Photo - 81874 - ARTS 1220 - 002 Basic Black & White Photo - 81875 - ARTS 1220 - 003 History/Appreciation of Art I - 81180 - ARTS 1310 - 001 : Our **sed** should have successfully stripped off the leading HTML text that we're uninterested in. Once that happens, suddenly our data becomes that much more readable. Note that there's still HTML data trailing our information. That can be addressed in another **sed** call: lab46:~$ cat fall2013-20110417.html | grep 'REGEX' | sed 's/PATTERN//g' | sed 's/<\/a>.*$//g' | less Accounting Practices - 81559 - ACCT 1000 - 001 Accounting Practices - 82350 - ACCT 1000 - 003 Financial Accounting - 82355 - ACCT 1030 - 001 Financial Accounting - 81558 - ACCT 1030 - 002 Financial Accounting - 81107 - ACCT 1030 - 003 Financial Accounting - 81108 - ACCT 1030 - 004 Financial Accounting - 81173 - ACCT 1030 - 005 Financial Accounting - 82115 - ACCT 1030 - 006 Managerial Accounting - 82078 - ACCT 1040 - 003 Accounting Procedures - 81123 - ACCT 1050 - 001 Federal Income Tax - 81783 - ACCT 1100 - 001 Federal Income Tax - 82358 - ACCT 1100 - 002 Intermediate Accounting I - 81124 - ACCT 2030 - 001 Intermediate Accounting I - 82359 - ACCT 2030 - 002 Computerized Accounting - 82361 - ACCT 2100 - 001 Cultural Anthropology - 81139 - ANTH 2120 - 001 Elem Mod Stand Arabic Con&StrI - 82296 - ARAB 1010 - 001 Elem Mod Arabic Con&StrI - lab - 82297 - ARAB 1010 - 071 Introduction Art Appreciation - 81505 - ARTS 1004 - 002 Drawing I - 81771 - ARTS 1030 - 001 Drawing I - 82176 - ARTS 1030 - 002 Drawing I - 82112 - ARTS 1030 - 003 Drawing I - 81503 - ARTS 1030 - 004 Ceramics I - 81151 - ARTS 1210 - 001 Ceramics I - 82110 - ARTS 1210 - 002 Ceramics I - 81504 - ARTS 1210 - 003 Ceramics I - 82134 - ARTS 1210 - 004 Ceramics I - 81176 - ARTS 1210 - 005 Basic Black & White Photo - 81873 - ARTS 1220 - 001 Basic Black & White Photo - 81874 - ARTS 1220 - 002 Basic Black & White Photo - 81875 - ARTS 1220 - 003 History/Appreciation of Art I - 81180 - ARTS 1310 - 001 In the provided expression, the following happens: * The pattern **<\/A>.*$** explicitly matches the closing "**a**" tag, and then matches whatever follows until the end of the line. * We replace that matched pattern with NOTHING. Note the presence of the backslash **\** before the closing slash of the **A** tag. This is needed because the forward slash **/** is the default field separator in **sed**(**1**), and to avoid the error of prematurely terminating the field, we use the backslash to escape it in order to match a literal forward slash. The result should be as appears in the sample above... no HTML data, just real readable course information. ^ 5. ^|Perform some data mining for me:| | ^ a.|Of this list, how many courses is CCC offering next semester?| |:::^ b.|How did you produce this result?| |:::^ c.|How many **CSCS** classes is CCC offering next semester? How did you find this?| |:::^ d.|How did you produce this result?| |:::^ e.|How many upper level (2000 and above) **ENGL** classes are being offered next semester?| |:::^ f.|How did you produce this result?| |:::^ g.|Who is offering more courses next semester, the English or Math department?| |:::^ h.|How did you produce this result?| Hopefully you're starting to see the value in what the Regular Expressions have enabled for us-- we were able to take raw data in some arbitrary format, and through analyzing it, adequately whittle away at it until it becomes a format readable to us. Once in that format, we can then perform some more valuable tasks on that data. =====Data Analysis===== In the **courselist/** subdirectory of the UNIX Public Directory are some additional files of value: * fall2010-20100315.html.gz * fall2010-20101113.html.gz * fall2011-20110417.html.gz * fall2013-20130417.html.gz * spring2010-20091022.html.gz * spring2010-20101113.html.gz * spring2011-20101105.html.hz * spring2011-20101113.html.gz * winter2011-20101113.html.gz Each of these files contains a snapshot of semester course information, referenced by semester, and snapshot date. Please make a copy of these additional files, uncompress them, and let's create a script to perform some data analysis. ^ 6. ^|Write a script that does the following:| | ^ a.|Accepts 1 or more of these files as an argument.| |:::^ b.|If no files are specified, display an error with usage information and exit.| |:::^ c.|If one file is given, perform the logic we've done manually on the command-line to produce and display the total number of courses offered in the given semester's course file.| |:::^ d.|If two files are given, and are both for the same semester+year, display the totals for each semester, and if the numbers do not match, display how both files differ (in an attempt to show what change took place).| |:::^ e.|If two files are given, and are **not** the same semester+year, display the totals for each semester, and display how many English courses are being offered in each of the files.| |:::^ f.|If more than two semesters are listed, do the same, but **also** display the totals for MATH, CSCS, BIOL, and PFIT.| |:::^ g.|Provide a copy of your script.| ^ 7. ^|As you are playing with the different course data files:| | ^ a.|Comparing fall2011 to fall2013, which semester offered more courses?| |:::^ b.|Do any of the files seem to break your logic?| |:::^ c.|Which one(s)?| |:::^ d.|Comparing a "working" file to a "nonworking" one, what seems to be a difference that trips up your patterns?| |:::^ e.|Between which two snapshot dates did this change seem to take place?| |:::^ f.|What can you surmise as being a cause of this change?| |:::^ g.|Could you adapt your script to handle the two different formats of data? How would you do this?| |:::^ h.|Provide a copy of your updated script.| __Hint:__ to compare differences between textual data sets, explore the **diff**(**1**) tool. There are a ton of questions we could ask of this data: * How many remedial (courses below the 1000 level) are offered a given semester? * Do any quantity of particular course(s) increase/decrease over time? * Is there a noticeable change in certain course offerings between a fall and a spring? Then there are some questions that, with our current skill level, may cause us a bit of trouble: * What is the range of CRN numbers for a given semester? (Lowest through highest) * Which course prefix has the MOST offerings a given semester? * Which course prefix has the LEAST offerings a given semester? * Which course prefix offered the MOST remedial course offerings? While we may be able to derive answers to these questions... in some respects the data is not conveniently arranged for our analysis purposes. At the moment we have our data in the following format: ^ UNIX/Linux Fundamentals - 81769 - CSCS 1730 - 001 ^ And to answer some of these questions, especially when **grep**'s are concerned, we'd ideally want the data arranged more like: ^ 78400:CSCS 1730-001:UNIX/Linux Fundamentals | So once again our data may not be exactly the way we want it. Do we give up? **HECK NO**, we conform the universe to our demands... =====Rearranging Data with Regular Expressions===== I consider where we are at now to be amongst some of the most powerful of concepts we learn in this class. What we are going to do now hopefully should take the cake and illustrate the true potential of the capabilities we are able to wield provided a good working knowledge of Regular Expressions and related tools. To do our next trick, we need to study our data once again: ^ UNIX/Linux Fundamentals - 81769 - CSCS 1730 - 001 ^ As you can see, the information as it is currently formatted takes the following structure, as compared to the desired structure: ^ Current: | Course Title - CRN - Course Prefix/Number - Section | ^ Desired: | CRN:Course Prefix/Number-Section:Course Title | So how could we do this? To accomplish this task, we need to explore another RegEx capability and exercise our options in the **sed** REPLACEMENT field. ^ 8. ^|With our data in the current structure:| | ^ a.|Derive a RegEx pattern that will match up to the first "space dash space". What is your pattern?| |:::^ b.|Derive a RegEx pattern that will match the CRN up to the second "space dash space". What is your pattern?| |:::^ c.|Derive a RegEx pattern that will match the Course Prefix/Number up to the third "space dash space". What is your pattern?| |:::^ d.|Finally, round out with a fourth RegEx pattern that matches the **Section**, which is at the end of the line. What is your pattern?| For my examples, I'll name your patterns REGEX1, REGEX2, REGEX3, and REGEX4. In order to rearrange our data, we need to effectively describe the data (as you did above) in order to reference it in groups. The RegEx symbols **\(** and **\)** denote Regular Expression groups, which we can use to isolate specific patterns for later reference. Check this out: lab46:~$ cat fall2013-20110417.html | grep 'REGEX' | sed 's/PATTERN//g' | sed 's/<\/a>.*$//g' > output lab46:~$ Notice what we just did here... we took our information in its current form of filtering and output it to a file (called **output**), effectively taking a snapshot of our progress. That should make sense, we're just using **I/O Redirection** to send the output of that pipelined command-line to a file instead of to **STDOUT**. Feel free to make use of similar output junctures during the solution of a problem like this- and who knows, you might need to do particular processing with certain arrangements of data. So if you output your data at certain key points, you could be making your work a lot easier. Moving on: lab46:~$ cat output | sed 's/^\(REGEX1\) - \(REGEX2\) - \(REGEX3\) - \(REGEX4\)$/\2:\3-\4:\1/g' | less 81559:ACCT 1000-001:Accounting Practices 82350:ACCT 1000-003:Accounting Practices 82355:ACCT 1030-001:Financial Accounting 81558:ACCT 1030-002:Financial Accounting 81107:ACCT 1030-003:Financial Accounting 81108:ACCT 1030-004:Financial Accounting 81173:ACCT 1030-005:Financial Accounting 82115:ACCT 1030-006:Financial Accounting 82078:ACCT 1040-003:Managerial Accounting 81123:ACCT 1050-001:Accounting Procedures 81783:ACCT 1100-001:Federal Income Tax 82358:ACCT 1100-002:Federal Income Tax 81124:ACCT 2030-001:Intermediate Accounting I 82359:ACCT 2030-002:Intermediate Accounting I 82361:ACCT 2100-001:Computerized Accounting 81139:ANTH 2120-001:Cultural Anthropology 82296:ARAB 1010-001:Elem Mod Stand Arabic Con&StrI 82297:ARAB 1010-071:Elem Mod Arabic Con&StrI - lab 81505:ARTS 1004-002:Introduction Art Appreciation 81771:ARTS 1030-001:Drawing I 82176:ARTS 1030-002:Drawing I 82112:ARTS 1030-003:Drawing I 81503:ARTS 1030-004:Drawing I 81151:ARTS 1210-001:Ceramics I 82110:ARTS 1210-002:Ceramics I 81504:ARTS 1210-003:Ceramics I 82134:ARTS 1210-004:Ceramics I 81176:ARTS 1210-005:Ceramics I 81873:ARTS 1220-001:Basic Black & White Photo 81874:ARTS 1220-002:Basic Black & White Photo 81875:ARTS 1220-003:Basic Black & White Photo 81180:ARTS 1310-001:History/Appreciation of Art I **__NOTE:__** If the format of your data does not seem to change, you've got a typo, or a RegEx that doesn't adequately describe the data. Go over your syntax, look for any possible gotchas. Ask questions, seek clarification, **and don't be afraid to have someone look at your pattern**... you'd be amazed what a second pair of eyes can do. Once you get it--- **WOW!** the data changed, just the way we wanted. Now we can do further analysis and write shell scripts that better assist us in our tasks. **Activities like this is what separates someone who can effectively command technology as a tool to assist you** to someone who resorts to manual data entry, racking up hours of time manually preparing the data to attempt to answer the same questions we've asked and gotten answers to. And our processing takes a fraction of the time it would take compared to trying to do all this data filtering and rearranging by hand. That is the power of Regular Expressions. We can effectively delegate the manual labor to the computer, which is very good at manual (and menial) tasks, and is great at following instructions. Plus, the less we are involved at the grunt-work level, the less chance there are of errors being introduced. The computer, when it follows correct instructions, will process the data effectively, versus the unpredictability of a human manually working on the data, accidentally inserting typos or other glitches that would threaten the validity of the end data. =====Additional Data Wrangling===== To cap off our experience, let's do one last foray into rearranging our data. ^ 9. ^|Rearrange the course information as follows (and show your command-lines):| | ^ a.|PREFIX NUMBER-SECTION(CRN):TITLE| |:::^ b.|PREFIX NUMBER:CRN (omit the section and title)| |:::^ c.|PREFIXNUMBER-SECTION:TITLE (CRN) (merge PREFIX and NUMBER together, no space separating them).| **PLEASE- ASK QUESTIONS, SEEK CLARIFICATION**. You're all just starting out, developing a proficiency with Regular Expressions. Typos happen. Don't let them trainwreck your progress on the lab. =====Conclusions===== This assignment has activities which you should tend to. While there is no formal submission required, it would be prudent to document some of the knowledge and experience gained on your Opus. As always, the class mailing list and class IRC channel are available for assistance, but not answers.