Skip to main content

Appending a stream of repeated bytes to IO in Python

I am writing a program that allows recording audio packets continuously from an event loop. Each iteration I would either get an actual packet of data, or no packet at all. I have to track the time since the last packet and append the silence packets to IO upon the next received packet.

The problem with that though is that the silence could be very long, and I don't want to create a large buffer of repeated data in order to write it to file, but it would also be extremely slow to write each silence frame individually in a loop, so my current solution is to combine the two and write in a loop with bigger to smaller segments of pre-generated silence.

I'm looking for a better optimal method. I have been looking at BufferedWriter's write method and noticed that it takes a bytes-like object. If that is possible, the optimal method would be being able to create a "bytes-like" object that simulates a stream of repeated data, that can be written x amount of times by the buffer without using a Python loop.

So far to no avail, so I'm asking this question here.

In order to simplify the question, I will be using representations of the data. buffer is always the returned result of open(file_name, "a+")

Suppose a silence packet is represented as "0", but of course in reality it is a lot larger than that, and 50,000 frames have passed since the last packet, I will need to append this silence to IO before writing the data from the packet. There are 3 ways of doing this I've thought of thus far.

The first way is to generate the data in memory first, then write it (which uses a lot of memory)

silence = b"0" * 50_000
buffer.write(silence)

The second way is to write all the silence frames in a loop (which is extremely slow)

for _ in range(50_000):
    buffer.write(b"0")

Lastly the third way is to combine the two, and write bigger segments (which is a lot faster than the second option, but slower than the first, and on top of that, it uses memory even when it's in use, plus overall this design is extremely ugly)

silence_1 = b"0"
silence_10 = b"0" * 10
silence_100 = b"0" * 100
silence_1000 = b"0" * 1000

# writing in a loop using bigger to smaller segments
# until there is no more silence to write
num_left = 50_000
while num_left > 0:
    if num_left >= 1000:
        buffer.write(silence_1000)
        num_left -= 1000
    elif num_left >= 100:
        buffer.write(silence_100)
        num_left -= 100
    elif num_left >= 10:
        buffer.write(silence_10)
        num_left -= 10
    else:
        buffer.write(silence_1)
        num_left -= 1


source https://stackoverflow.com/questions/77462225/appending-a-stream-of-repeated-bytes-to-io-in-python

Comments

Popular posts from this blog

ValueError: X has 10 features, but LinearRegression is expecting 1 features as input

So, I am trying to predict the model but its throwing error like it has 10 features but it expacts only 1. So I am confused can anyone help me with it? more importantly its not working for me when my friend runs it. It works perfectly fine dose anyone know the reason about it? cv = KFold(n_splits = 10) all_loss = [] for i in range(9): # 1st for loop over polynomial orders poly_order = i X_train = make_polynomial(x, poly_order) loss_at_order = [] # initiate a set to collect loss for CV for train_index, test_index in cv.split(X_train): print('TRAIN:', train_index, 'TEST:', test_index) X_train_cv, X_test_cv = X_train[train_index], X_test[test_index] t_train_cv, t_test_cv = t[train_index], t[test_index] reg.fit(X_train_cv, t_train_cv) loss_at_order.append(np.mean((t_test_cv - reg.predict(X_test_cv))**2)) # collect loss at fold all_loss.append(np.mean(loss_at_order)) # collect loss at order plt.plot(np.log(al...

Sorting large arrays of big numeric stings

I was solving bigSorting() problem from hackerrank: Consider an array of numeric strings where each string is a positive number with anywhere from to digits. Sort the array's elements in non-decreasing, or ascending order of their integer values and return the sorted array. I know it works as follows: def bigSorting(unsorted): return sorted(unsorted, key=int) But I didnt guess this approach earlier. Initially I tried below: def bigSorting(unsorted): int_unsorted = [int(i) for i in unsorted] int_sorted = sorted(int_unsorted) return [str(i) for i in int_sorted] However, for some of the test cases, it was showing time limit exceeded. Why is it so? PS: I dont know exactly what those test cases were as hacker rank does not reveal all test cases. source https://stackoverflow.com/questions/73007397/sorting-large-arrays-of-big-numeric-stings

How to load Javascript with imported modules?

I am trying to import modules from tensorflowjs, and below is my code. test.html <!DOCTYPE html> <html lang="en"> <head> <meta charset="UTF-8"> <title>Document</title </head> <body> <script src="https://cdn.jsdelivr.net/npm/@tensorflow/tfjs@2.0.0/dist/tf.min.js"></script> <script type="module" src="./test.js"></script> </body> </html> test.js import * as tf from "./node_modules/@tensorflow/tfjs"; import {loadGraphModel} from "./node_modules/@tensorflow/tfjs-converter"; const MODEL_URL = './model.json'; const model = await loadGraphModel(MODEL_URL); const cat = document.getElementById('cat'); model.execute(tf.browser.fromPixels(cat)); Besides, I run the server using python -m http.server in my command prompt(Windows 10), and this is the error prompt in the console log of my browser: Failed to loa...