it started as a joke on postmodern theory, got accepted in academic articles. Later same happened with computer science and mathematical articles. -> shows how language functions in specific social settings,
look at algorithms, see how they work
computed generated erotic texts--can't tell (stereotypical tone, more easy to generate)
study "grammar", looking at code
everything predefine, nevertheless seems human made..(=realistic?)
algorithm learns to speak (shakespeare)-->feeding a lot of text, build math model, optimizing
www.youtube.com/watch?v=qv6UVOQ0F44 computer learning winning/playing video games
masked "types" like in commedia dell'arte
------------- starting from tools, start playing explore tools in different ways (don't need to know programming)
computer generated bullshit
text-mining/data-mining-->can use pattern library too
"the annotating has been done long time ago..by shakespeare"
responding to computer generated spam
/ an inventory of experiences
"they looked at the website, they didn't read it" nuclear w. parody. exploiting ritualized language situations
/ an inventory of absences
/ a collective glossary
ritualized language situation
/ an inventory of questions
/ an inventory of tensions
********************************************** 21st August round of presentation/ideas:
Gijs: Generate texts from / for politicians
Anna: From Barcelona, working with code for interactive installations. Interested in the track to be able to generate content which can pass a turing test. Also can feed it to other tracks? Can we generate text for politicians based on "langue de bois" text? Can this text be used in politics to get to an agreement? Therefore, can we replace politicians by politic bots?
Ricardo from Porto : interested in text-generation with a Dada-ist approach. Emulating interest ; create system to replicate, rather than passing Turing test like humans. Also worked with automated lay-outing these generated texts to turn them into objects.
Ana Isabel: also from Porto, designer. Collaborating with Ricardo. Has been experimenting with text. Have Portugese material, also other material they can share. Fado lyrics, texts from Parlement, also done experiments with newstitles.
Wembo: Graphic designer, previous project was a Prejudice Generator (executing/speaking without thinking). Grammar structures for prejudice, local variations. (combination of different populations to make the generated text more and more absurd) Prejudice different from place to place, learning to be as absurd and offensive as human beings
Jara: Language research group, contemporary poetry, in Madrid. Worried about depolitization of the new conceptualists, and interested in tech approach of their work. Euraca (?) group. Present languaging. Mediation through technological layers, as well as contextual or oral ones. Trying to learn simple methods of text manipulation [euraca: about https://seminarioeuraca.wordpress.com/about/ // conceptualismos yankis: https://seminarioeuraca.wordpress.com/programa3/ ]
Raphael: curious to see how for we can go. If code can become text, can code generate code again. Perhaps not only code, but also SVG. How graphical can these generated results be? Clipart generator, after having the progran learn about it from OpenClipArt. A program that learns about shapes and can generate them raw semantic, mark down
Hans: diverse background. Little bit of programming, little bit statistics. Use the program as a political activist.
Catherine: get more familiar with the text generator as a writing and reading machine. Take it as a political tool based on a collection of texts. Same interest in the “langue de bois” discourses on Text & Data Mining Illegaly accessing sources of copyrighted texts for data-mining (there are exceptions for education purposes)
Samuel, graphic designer. Worked on a chat bot, based on DB. Interested about how Neural Networks are done. interest in markov chains, neural networks,
Andrea : studied graphic design but drifted away from it, interest in language - topical research -- interest on ritualistic aspect of language, experimental poetry, constructed as if was made by a machine. Interest in poetry made as if it would be made by machines, but actually made by humans. Interested in this shifts.
Two methods/approaches to work with text: - sintax and language-based, the most easy to get in - neural networks (most challeging)
Dada engine, a C program http://dev.null.org/dadaengine/ Does’t need to get into the code But you need to compile the package. Look at the README-file: after extracting you need to go into the dada-engine folder and to do 'configure', 'make' and then 'sudo make install' (The sudo is needed to have the permission to create a directory). Probably you get errors and have first to install the following files: bison and flex (with 'sudo apt-get install bison' and 'sudo apt-get install flex'.) people working with Ubuntu need to install also makeinfo (sudo apt-get install texinfo)
Tests to generate some text (example): dada scripts/manifesto.pb The NonIsmist Manifesto: 1. language is an illusion. 2. culture is a myth.
The ParaNihilist Manifesto: 1. there is no recovery. 2. reality is a myth.
The NonSurObjectivist Manifesto: 1. class is an illusion. 2. technology is a myth.
Ricardo/Ana: We feed dasher with a corpus of Fado, portuguese music lyrics -> [[Dasherpoetry]] Working on alternative input for the interface to generate text. Possible corpus to try: fado lyrics; portuguese parliament debates; recipes; code (tried it with kernel.c);
Ann: Installing dada engine
Hans: installed neural network library; now feeding it
Raphael: running minimal RNN python library to generate SVG
Data mining (the analysis step of the "Knowledge Discovery in Databases" process, or KDD), an interdisciplinary subfield of computer science, is the computational process of discovering patterns in large data sets ("big data") involving methods at the intersection of artificial intelligence, machine learning, statistics, and database systems. The overall goal of the data mining process is to extract information from a data set and transform it into an understandable structure for further use.
Get the data: A json dump of the database is available at http://10.9.8.57/export You can download it with the following command wget 10.9.8.57/export -o export.json Which will download the data into a file called 'export.json' in your current directory.
It's a list of object / dictionaries with the keys
url => the url of the page from which the speech/transscript was downloaded text => the origninal text (only the parts spoken by Obama) cleaned_text => a (roughly) cleaned version where double spaces and annotations like (applause) & (laughter) have been removed
Interview machine More an idea than a finished product yet. Basic idea is to use the Dada Engine for its structured templates, but make it more open by creating such pb-files with scraped information. At the moment the only thing it does is scraping headlines from Reuters and posing a question. This can be enlarged with an answer module based on the article itself. How exactly to create the answer raises a lot of possible options from simple pre-cooked structures to using more advanced methods using semantic ontologies. ../text-generation/interview%20machine.py