Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Update icon and title audio classification #38

Open
wants to merge 2 commits into
base: main
Choose a base branch
from

Conversation

st-tuanmai
Copy link
Contributor

No description provided.

@PaulTR
Copy link
Collaborator

PaulTR commented Oct 9, 2024

Can you verify the classification happening in the app? It's picking up a lot of random things that aren't happening, and not picking up things like "speech" or "whistling" when that is happening. Thanks.

@st-tuanmai
Copy link
Contributor Author

Can you verify the classification happening in the app? It's picking up a lot of random things that aren't happening, and not picking up things like "speech" or "whistling" when that is happening. Thanks.

I will check it now

@st-tuanmai
Copy link
Contributor Author

I updated the branch to change TensorFlowLiteTaskAudio to TensorFlowLiteSwift.
Please help me check again.
Thanks you.

@PaulTR
Copy link
Collaborator

PaulTR commented Oct 28, 2024

I am still seeing wrong results consistently with this. Just whistling into the phone and not getting 'whistling'.

@st-tuanmai
Copy link
Contributor Author

Hi @PaulTR
I checked the model label file and the speech_commands label doesn't have whistling and speech. Please help me check.

labels:
background
down
go
left
off
on
right
stop
up

@PaulTR
Copy link
Collaborator

PaulTR commented Oct 29, 2024 via email

@PaulTR
Copy link
Collaborator

PaulTR commented Oct 29, 2024 via email

@st-tuanmai
Copy link
Contributor Author

I am using label.txt file extracted from tflite file using python code, I think label is added wrongly to metadata.
vocal = zipfile.ZipFile('./demo/speech_commands.tflite').extractall('label')

@PaulTR
Copy link
Collaborator

PaulTR commented Oct 30, 2024

Not comparing the speech commands model - can you verify this all with the standard audio classification model (the one with 'music', 'whispering', 'whistling', etc.)? Thanks.

@st-tuanmai
Copy link
Contributor Author

st-tuanmai commented Nov 1, 2024

@PaulTR
Copy link
Collaborator

PaulTR commented Nov 1, 2024

OK again we're not looking at the speech models for this issue. Only looking at the sound classification model that does whistling/whispering/music/etc. the label that shows up is not matching the sound.

Delete the speech model if that's adding a complication, it doesn't matter for the sample. We need the first general sound classification model to work correctly.

@PaulTR
Copy link
Collaborator

PaulTR commented Nov 5, 2024

You can compare the iOS sample to the Android sample. They're using this model: https://storage.googleapis.com/ai-edge/interpreter-samples/audio_classification/android/yamnet.tflite (I copied the same one into the ios folder) and returning the correct results.

@st-tuanmai
Copy link
Contributor Author

@PaulTR
Copy link
Collaborator

PaulTR commented Nov 6, 2024 via email

@st-tuanmai
Copy link
Contributor Author

@PaulTR I removed the speech commands model, I think we can use only yamnet model now.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants