Guitar

I bought a cheap guitar in April and have been teaching myself how to play since. Although it’s not particularly avant-garde of me, it is always valuable to learn new things and it has already given me a different perspective as a non-musician.

Playing an instrument takes time. Your fingers hurt, taking mental and physical fortitude. I have so far found it rewarding

The way chords and notes are approached is interesting. The 6 strings mean 6 possible notes a chord, introducing the concept of drone chords, and lower pitch strings allowing for a structural base note that progresses down the line. 

Standard and non-standard tunings, open-scales and Nick Drake. Using a chord as a basis for a scale.

I chose to try to learn the guitar for a number of reasons:

  • I wanted a physical instrument as an audio source to process effects through (mostly midrange frequencies)
  • I’m a fan of guitar music (particularly in the context of shoegaze and earlier post-rock)
  • My dad plays the guitar, allowing him to teach me some stuff as well as a bonding activity 🙂

I am a fan of playing instruments unconventionally. I have previously read about extended techniques when it comes to guitar. Earlier in the year, I read about Keith Rowe and the techniques he used. I also read about Glenn Branca who I did not know about before, even though he is very influential. I sometimes like bringing my phone up to the pickups to get electrical interference and then using it as a slide.

The more ambient pieces below were made while having the guitar flat on my bed and occasionally lightly tapping the strings with a flatlined compressor and a bunch of effects in Ableton (Valhalla delay/vintage verb, pedal effect and the same max effect I’ve been using a lot recently).

Here are some examples of a recent improvised/experiment/practice with just guitar and effects (was recorded live in one take so its not mixed all that great):

Don’t feel obliged to listen to them all, I just had to break them up due to the file size limit.

More Ambient

More Guitar, a bit gloomy

An accidental mess of a combination of all these files on top of each other when exported

As is sometimes the way, on listening back to these audio clips, it sounded better at the time of playing/recording it. I could somewhat blame it on these files being compressed mp3s so that I could actually upload them due to the 1o mb upload limit.

Generative music

Generative music – designing a system to create music

I watched an online interview/class with Mark Fell. 

https://en.wikipedia.org/wiki/Snd_(band)

https://www.youtube.com/watch?v=EcTAAdDPhTk.

https://www.thewire.co.uk/in-writing/essays/collateral-damage-mark-fell

He talks about Max patches as compositions in themselves. I’ve had similar thoughts recently on the format of digital audio software and daws. I was thinking about the idea of a music listening platform that only plays ‘patches’, allowing others to listen to music but also see what’s underneath it (how it was made and how it performs itself)

Laurie Spiegel

https://teropa.info/musicmouse/

“I automate whatever can be automated to be freer to focus on those aspects of music that can’t be automated. The challenge is to figure out which is which.”

Brian Eno

Steve Reich

Markov chain

65 days of static z03

65DOS also made the score to No Mans Sky, games and generative music go hand in hand.

https://en.wikipedia.org/wiki/Tree_(data_structure)

Sound For Screen

Today we improvised over film clips using hardware synthesizers (Korg Volcas). I mostly used the Volca Kick and the Volca keys.

Sound and Music in film are very important. It can manipulate emotions and completely alter the meaning of scenes. I always think back to an example I was shown doing Media at school of scenes from horror films with different sounds on top of the video.

Initially, I found the limitations of the Kick quite restricting, mostly as I was unfamiliar with using this specific device, however, as I experimented further I found that the drive and envelope could be used to get more effective sounds, especially in the context of the clip we were scoring to from Bladerunner.

Many people in class found that as they rewatched the clip they found their improvisation to be less sporadic and not as effective. I found it to be the opposite as after rewatching the clips I had a better understanding of the rhythm of the editing. It also allowed me to become more familiar with the instruments at the same time, seeing what worked and what didn’t while experimenting.

I think one of the most important elements this introduced was the use of silence, and how when being told to improve over the clip as a single performer it’s important to resist the urge to fill the entire timeframe with sound. It makes me think back to my previous comments on John Cage, as well as on Talk Talk’s Spirit of Eden / Laughing Stock and post-rock in general. I think that if we’d had heard everyone’s sounds at the same it would have been an interesting exercise

I mostly found myself creating more musical soundscapes for the clips, as opposed to dietetic sounds. I enjoyed watching the clips with their sound after our improvised attempts. I found Johnny Greenwood’s score, and the use of audio in You Were Never Really Here to be very effective, especially in how it interacted with the scenes, placing the sound as a focus in its creation as opposed to just a tacked-on feature afterwards or for the sake of sensory continuity.

Something I found interesting is that due to being familiar with the existing scores of the films shown, it was somewhat distracting not to mimic them. This helped in the case of Vangelis’s score as I was using the Kick so it wasn’t really possible to copy.

Here is the original improvisation with Key over You Were Never Really Here, as well as an unfinished creation using only the improvisation as material.

On a tangent, I personally have a large amount of respect for films with a very minimal soundtrack (if any at all). I’ve consistently mentioned it this year but Chantel Akerman’s ‘D’Est’ is one of my favourite films as it is essentially field recordings with visuals. There is no story, just video and sound yet to me each scene tells its own story.

Controllerism

  • Using phones, tablets, and computers to manipulate recorded elements in DAWs 
  • Investigating iOS and Android hardware potential for sound manipulation e.g. accelerometer, gyroscope 
  • Apps, such as L.K. and Touch OSC, for controlling Ableton (iOS and Android) 
  • Mapping phones to computer DAWs.  Routing iOS keyboards in desktop synths (expressionPad MIDI/Synth – freeware) 
  • Creating custom layouts for iOS and Android (e.g. Touch OSC) Using iPhone/ Android gyroscope to generate midi data

Today we looked at controllerism, focusing on using Ableton Live’s midi mapping.

Since getting an Ableton Push 2 I seldom find myself using Live without it. I enjoy the Simpler controllers for sampling as well as the Grid layout which makes it very accessible for someone like myself who doesn’t play keys and would otherwise get stuck in the same few scales.

The Push 2 has polyphonic aftertouch and pressure but does not have full MPE features such as slide. The Wavetable synth inside Ableton is the best for exploring this as it fully supports MPE meaning you can change the wavetable position for each note you’re playing depending on the pressure used (with each finger) allowing for more control over sounds.

Instead of using the Midi Mapping feature I often just make device racks that automatically come up on the push, allowing me to make my own instruments and presets.

Earlier in the year, I experimented with a VR headset and found it to be an exciting field to explore due to the specialisation of sound and the possibilities of interactability and controllerism. Using unity I created objects that emitted sound and could physically interact with them, moving them closer or further away. Some of these objects had physics and would fall to the ground, while others were frozen in space when let go of. This is something I would like to explore in the future, exploring how techniques made accessible through VR could be used in an artistic way as opposed to just novelty.

My laptop also has a touchscreen, something that I rarely use but have once before while experimenting with a Max patch. (as seen below) For my previous project, I mapped a foot pedal to Max so that I could control the length of the loops while interacting with a microphone, allowing me to be away from the computer itself.

https://photos.app.goo.gl/9GRgPLZzrnmLe1Er6

I have a small groove box/sampler/midi sequencer/controller called the OP-Z which has a built gyroscope for modulation yet I seldom use the feature. I enjoy using the devices as a sequencer and a tool for making music however I still have mixed feelings on it as a whole as for the same price I could have got a physical synth that would work alongside Ableton and Max.

I find it unlikely at the moment for me to use my phone as a controller.

For my hand-in, I aim to create a generative piece to explore creating a piece with no interaction on my part, other than pressing play/record.

Homework:

1. Develop one of the pieces begun in today’s session to be played in next week’s session.

Creative Synthesis

P1 – Software

We played around with subtractive software synths in different DAWs. The one in Logic (retro synth) sounded more like a hardware synth, with the filters sounding more analogue whereas the Ableton Analog instrument sounds less thick, and the interface is a bit of a mess. My personal favourite Ableton synths at mimicking analog devices are the POLI and BASS Max devices.

We also tried using VCV rack. I have a small amount of experience in VCV as I have tried it before. VCV rack simulates a modular set-up very well and can produce very nice sounds. I find it appealing that it is free to use and that there are so many free modules to use. The main downside is that you don’t get the physical experience of patching like you would with a real Eurorack modular setup, however, the fact it is free more than makes up for this (especially considering the price of an actual modular setup).

Here are some recordings I made of POLI and BASS in Ableton:

Advantages of software synths:

  • Always Midi compatible meaning you can rearrange your input after ‘recording’
  • Cost/Accessibility/Practicality – Software synths are generally cheaper and cost less power to run. As well as cost, they take up no physical space and don’t need to be set up with cables and external systems that could otherwise cause complications (input latency etc.)

Disadvantages

  • Sound Quality – The quality can vary between software synths and they often don’t sound as good as analogue
  • Tactility (controller dependant) – A big part of using a synth is the facility of interacting physically with it while using it.
  • Character – This is as much a case-by-case (negative and positive) and personal opinion but due to the precise reliable nature of a lot of software it misses out on the quirks of individual devices such as hardware errors.

I created this basic synth in Max to try to get a better understanding of how FM works. Due warning, the audio below is quite coarse and loud so try to start with low volume.

Some examples of the synth above. No effects. Just a pentatonic arp.

I currently use software synths as my main source of synthesis. I find it accessible and easy to understand. I hope to learn how to make my own complex synths in Max/MSP. I would like to get into hardware synths however my budget is holding me back, as well as the choice of many different synths. I was looking at the following on Thomann a while back: ASM Hydrasynth Explorer, Dreadbox Nymphes, Korg Wavestate

I created the soundscape below in Ableton. 1 track was an Operator (the filtered wind sound), the second track was processed recordings of a radio station (one of those beeping shortwave ones like the pip – audible parallels with Bill Fontana’s Fog Horns) and the 3rd track is electric guitar filtered very low. It’s too large a file to attach so here is a link that should hopefully work.

https://drive.google.com/file/d/1zUuue2_Ad8OzKmlTHOwv0v3RnikfF-xS/view?usp=sharing

P2- Hardware

Today we used the hardware synths available at LCC.

My personal favourite out of these was the Moog Matriarch as it sounded the best and I enjoyed switching between the different voicing settings. Of all the synths available at the time it was probably the only one that appealed to me more than just using familiar software.

When I first arrived on the course my knowledge of synthesis was self-taught based on using software synths in daws and memories from the very cheap Casio keyboard at my Nans. Although I knew how ADSR envelopes and filters functioned it was initially quite daunting trying to see how they were arranged on specific synths, like the MS-20. Thankfully I’m now a lot more familiar with hardware design conventions.

Do you agree all the points made here? What has been left out?

Music Concrete

History of music Concrète – notable figures etc.

Pierre Henry, Pierre Schaeffer, Daphne Oram, Delia Derbyshire

In today’s session, we used Ableton to manipulate samples that we recorded around the uni building. I would say that I’m very familiar with sampling in Ableton as I often resample sounds that I have previously made and recorded. I often use the Simpler device with my Push 2 due to its intuitive visual interface and flexibility in playback (with the option to playback loops polyphonically, as well as chop the sample into 64 smaller slices). The Sampler device is better for multisampling, however, the interface is less intuitive on the Push.

I’ve read about Music Concrete before as I’m very interested in sampling, plunderphonics and the concept of sampling in relation to soundscapes such as KFL’s Chillout, the music of Boards of Canada and some of Lawrence English’s work. This is something I hope to explore further in the future, and will likely explore in the context of digital worlds due to my experience with game engine environments.


Inspired by Music Concrète, I made a piece. The sound sources are a paper booklet, me cracking my knuckles, an old radio drama recording, some cello samples and feedback sounds from playing headphones over a microphone.

I used Max to speed up sounds beyond recognition and used Paulstretch to slow down more percussive sounds to completely remove the transient. I also used max to create a looping effect that randomly changes the loop length (and playback rate of the sound). I arranged it in Ableton live 11.

I think I should have attempted to create more percussive ‘attack’ focused sounds. This juxtaposition would have created more variety in the piece and aided in the compositional structure to give a sense of tension.


The concept of sound objects makes me think of Matmos and their use of sampling and abstracting physical materials.

Types of Listening

Causal Listening: Listening for the purpose of gaining information about the sound’s source.

Semantic Listening: Listening for the purpose of gaining information about what is communicated in the sound.


Reduced Listening: Listening to the sound for its own sake, as a sound object, by removing its real or supposed source and the meaning it may convey.

Deep Listening: Listening to everything all the time, and reminding yourself when you’re not. Actively listening, both in the moment, and outside.