Shortening live sound for monitoring?

Discuss everything related to production, recording, composition, etc.

Moderators: lisa, Kent, Joe.

Post Reply
Spectromat
Common Wiggler
Posts: 119
Joined: Mon Oct 13, 2014 11:03 am
Location: UK

Shortening live sound for monitoring?

Post by Spectromat » Sun Oct 06, 2019 3:38 am

Hi all,
I am trying to figure out if there might be a way to monitor live sound in a way that cuts out part of the live sound (75% of the sound duration) allowing me to monitor a shortened version of what is playing (25%) in a meaningful way?

The idea stems from thinking about recording slow motion video whilst at the same time creating music/sound/noise that will eventually play back with the slow motion video, but I would like to be able to manipulate the sound devices in time with the camera movement (and over time stay in sync).

Not sure if I need to be thinking about software that in some way slices the sound and plays back using some sort of live granular synthesis, or if I need to create a tempo device that runs four times faster than the sounds being created? Also unsure if there might be any hardware solution that would be needed to pitch shift the live feed?

Any thoughts or ideas would be gratefully received :tu:

Koekepan
Wiggling with Experience
Posts: 492
Joined: Sun Nov 26, 2017 10:50 pm
Location: PNW

Post by Koekepan » Sun Oct 06, 2019 12:36 pm

I'm really not clear on your goal here.

Are you trying to cut, say, the tail of a snare during live performance?

Spectromat
Common Wiggler
Posts: 119
Joined: Mon Oct 13, 2014 11:03 am
Location: UK

Post by Spectromat » Sun Oct 06, 2019 2:12 pm

Apologies if it’s not very clear, that’s probably because I am not entirely sure anything like this is possible..

I would like to shoot some slow motion video (say 96fps for ease of calculation) whilst creating music as the video is shot, and I have been wondering if there is a way to monitor / record the music at the same time as the video is shot.. but to somehow slow down the live recorded music (slowed by 25% so it will play back with the footage at 24fps) so I can hear and use the sound cues to move the camera in time with the music being created?

Sorry if it sounds rather convoluted, but I have done a whole load of searching and not found any references to help me explain it better..

Hope that makes a little more sense?

Koekepan
Wiggling with Experience
Posts: 492
Joined: Sun Nov 26, 2017 10:50 pm
Location: PNW

Post by Koekepan » Sun Oct 06, 2019 8:48 pm

OK, so you're shooting something at 96fps with the intention of displaying it at 24fps.

So if you're playing at 240bpm, it will come out as 60bpm.

Let's assume that 60bpm ambient in time with flowing water in a creek for a meditation video is your goal.

Now you want to hear your 240bpm madness so that you can react to it in a timely fashion, as a director of photography?

I'm not about to say that this is impossible, but it seems unnecessarily difficult. Why not shoot footage first and synchronise later? If you have a solid beat to you rmusic, why not just play a metronome? I have the feeling there's an awful lot you're not telling us about this scenario, because as described the constraints sound weird as hell.

Spectromat
Common Wiggler
Posts: 119
Joined: Mon Oct 13, 2014 11:03 am
Location: UK

Post by Spectromat » Mon Oct 07, 2019 2:37 am

Koekepan wrote:I have the feeling there's an awful lot you're not telling us about this scenario, because as described the constraints sound weird as hell.
The reason for not getting into too much detail, is because I am currently exploring the potential of a situation where sound is recorded live, my background is as a visual artist, where having flexible scenarios can prevent one from missing potential creative avenues for a project's realisation, nothing more sinister than that.. 8-)

Art often creates these weird constraints, it is true, but sometimes these constrains can help to reveal something interesting, and I am exploring the potential around a moment of experience where visual and audio information are connected.

The scenario of flowing water filmed at 96fps (for 24fps playback) and 60bpm ambient sound played at 240bpm is well observed, and close to some of my recent visual experiments. I am currently exploring less beat driven sounds, with sample based ambient drones created through loop tweaking and granular manipulation and some analogue synth modulation, but maybe pattern based things could work.

Ultimately I am thinking of a camera rig / synthesizer rig that includes both image capture and sound generation / recording gear so that things like focus, zoom and camera movement can be connected to alterations of audio parameters.

Whether the output of this is for live performance or the recreation of specific events where image and audio are captured together for screening later still isn't that clear to me yet, thanks for your current suggestions :tu:

Koekepan
Wiggling with Experience
Posts: 492
Joined: Sun Nov 26, 2017 10:50 pm
Location: PNW

Post by Koekepan » Mon Oct 07, 2019 8:14 pm

Then what I would suggest to you, in the absence of a specific use case (or rather, more specific than capturing live sound) would be to be prepared to use pitch shifting so that the sounds aren't all redshifted out of existence, with perhaps a gating effect to trim audio tails where available.

Spectromat
Common Wiggler
Posts: 119
Joined: Mon Oct 13, 2014 11:03 am
Location: UK

Post by Spectromat » Wed Oct 09, 2019 12:41 pm

Thanks Koekepan :tu:

Appreciate any contributions :)

Technologear?
Wiggling with Experience
Posts: 271
Joined: Fri Aug 11, 2017 7:08 am
Location: Straya

Post by Technologear? » Wed Oct 09, 2019 4:53 pm

The music video guys will have lots more to say on this topic than most around these parts of the internet. Not sure where that hang out.
I thought of this:
[video][/video]
Get your ambient chipmunk on

Edit: sorry I can't embed links properly[/video]

Spectromat
Common Wiggler
Posts: 119
Joined: Mon Oct 13, 2014 11:03 am
Location: UK

Post by Spectromat » Wed Oct 09, 2019 5:47 pm

[video][/video]
That's a good point..

I initially thought this was going to be down to audio processing (to remove some of the audio signal) I wondered if something like max msp might be able to chop the audio up and feed parts of the chopped up sound into a mix running at a quarter of it's live speed.. but learning the language to code, without knowing if it might work could take a while.

Will check out some video hangouts and ask around 8-)

(to embed the video remove the s after http;)

Post Reply

Return to “Production Techniques”