Search Results

Search found 16362 results on 655 pages for 'audio interface'.

Page 109/655 | < Previous Page | 105 106 107 108 109 110 111 112 113 114 115 116  | Next Page >

  • UITableView section footer multiple UIButtons

    - by joec
    In the section footer of my UITableView, i have one UIButton, this button is 320 pixels wide (i.e. takes up full width of view). I would like two buttons in the section footer, but when i try and add another button, Interface Builder just puts it on top of the existing button. If i try to change the size in the inspector from 320 to say 50 pixels, it always reverts to 320. What am i doing wrong? Thanks.

    Read the article

  • +(void) initialize in objecive c class static variables construstor

    - by sugar
    I found some sample code from here. static UIImage *backgroundImageDepressed; /** * */ @implementation DecimalPointButton + (void) initialize { backgroundImageDepressed = [[UIImage imageNamed:@"decimalKeyDownBackground.png"] retain]; } is it something like this - +(void) initialize method initialize static variables of a class ( interface ) in objective c ? I have never seen this before. Please need your guidance on it. Thanks in advance for sharing your great knowledge. Sagar

    Read the article

  • creating a shared library from c++ source file

    - by navinbecse
    when I tried to create a shared library file using the "cl" command in the vc++ command prompt, it shows a error saying "Can't open include file "jni.h": No such file or directory"... the jni.h is tried to be included in the machine generated header file from java class... i am using this for java navite interface operations... can any one help me...

    Read the article

  • iOS: how to understand the role of rootviewcontroller and its relationship with other objects

    - by Anthony Kong
    I have created a UISplitViewController based iOS app in XCode 3.2.5 Below is a screen shot of Interface builder showing the rootviewcontroller and how it is linked to other objects. Being a beginner myself, I do not understand: 1) What is the role of the rootviewcontroller? Searched the documentation but what I found did not answer this question. 2) I thought a IBOutlet should only link to one corresponding object. Why in this case the rootviewcontroller is linked to two?

    Read the article

  • How to insert a WebView into a cocoa app?

    - by spamoom
    I'm rather new with the whole OSX programming, I wanted to stick a WebView into an empty app. Apparently it isn't as simple as sticking a WebView on a window in interface builder and creating an outlet. IBOutlet WebView *webView; It gives me a expected specifier-qualifier-list before 'WebView' and when I don't use an outlet, it terminates due to uncaught exception. I'm not too sure what these error messages mean. Seems it isn't that simple!

    Read the article

  • How does one create an elegant iPhone GUI?

    - by jrtc27
    This is just one of those things where you feel like your own design is utterly terrible, and that all of the other apps have a beautiful design. This question is just about how you would go about creating a user interface that a user would actually want to use?

    Read the article

  • IList<T> and IReadOnlyList<T>

    - by Safak Gür
    My problem is that I have a method that can take a collection as parameter that, Has a Count property Has an integer indexer (get-only) And I don't know what type should this parameter be. I would choose IList<T> before .NET 4.5 since there is no other indexable collection interface for this and arrays implement it, which is a big plus. But .NET 4.5 introduces the new IReadOnlyList<T> interface and I want my method to support that, too. How can I write this method to support both IList<T> and IReadOnlyList<T> without violating the basic principles like DRY? Can I convert IList<T> to IReadOnlyList<T> somehow in an overload? What is the way to go here? Edit: Daniel's answer gave me some pretty ideas, I guess I'll go with this: public void Do<T>(IList<T> collection) { DoInternal(collection, collection.Count, i => collection[i]); } public void Do<T>(IReadOnlyList<T> collection) { DoInternal(collection, collection.Count, i => collection[i]); } private void DoInternal<T>(IEnumerable<T> collection, int count, Func<int, T> indexer) { // Stuff } Or I could just accept a ReadOnlyList<T> and provide an helper like this: public static class CollectionEx { public static IReadOnlyList<T> AsReadOnly<T>(this IList<T> collection) { if (collection == null) throw new ArgumentNullException("collection"); return new ReadOnlyWrapper<T>(collection); } private sealed class ReadOnlyWrapper<T> : IReadOnlyList<T> { private readonly IList<T> _Source; public int Count { get { return _Source.Count; } } public T this[int index] { get { return _Source[index]; } } public ReadOnlyWrapper(IList<T> source) { _Source = source; } public IEnumerator<T> GetEnumerator() { return _Source.GetEnumerator(); } IEnumerator IEnumerable.GetEnumerator() { return GetEnumerator(); } } } Then I could call Do(array.AsReadOnly())

    Read the article

  • IBOutlets are always nil

    - by Jonas
    Hi, In IB I have subclassed the File Owner and the subclass loads my .nib. The subclass itself doesn't need access to the IBOutlets so they're instance variables in another class, which I added as a subclass of NSObject in the .nib. Then I connected the interface elements to that object. But somehow, in the class that "owns" the IBOutlets, they show up as nil. What am I doing wrong here?

    Read the article

  • The fastest way to encode image+audio for Youtube from command line?

    - by Pavel Vlasov
    I have an mp3 and image and I want to make a simple clip to upload onto Youtube. Is there a fast solution? If video formats are so bad designed, then maybe it is possible to use a prerendered video-only clip? This works good except it takes as much time as the audio lasts: ffmpeg -loop_input -r ntsc -i "%IMAGE%" -i "%AUDIO%" -r 1 -acodec copy -shortest -re -force_fps "%VIDEO%" This takes a second but results in a black screen video that is successfully played by a desktop video player but not acceptable by Youtube: ffmpeg -i "%IMAGE%" -i "%AUDIO%" -acodec copy "%VIDEO%" Windows 7. Preserving audio quality is preferred over video quality.

    Read the article

  • How do I merge MP4 files without audio going out of sync?

    - by djangofan
    Is there a tool I can use that can merge MP4 files without throwing the audio out of sync? I generated some MP4 files from a DVD using AVIDemux but whatever tool I try to use always ends up throwing the audio out of sync with the video. The further you get into the video the further off-sync the audio is. By themselves the MP4/AAC videos have perfect audio-video sync. later tonight i might try http://www.headbands.com/gspot/ to examine the file before and after to see if anything changed in the media format.

    Read the article

  • How do I add another audio stream to an MP4 file?

    - by RandomEngy
    I've got an MP4 video file and I want to add another AAC audio track to it. I've tried YAMB and MeGUI (frontends for MP4Box) and it plays correctly in Zoom Player, but it picks the wrong track in WMP and plays both at once in Quicktime. I think this might have to do with designating the default audio track somehow. Does anyone know how to specify the default audio track with YAMB/MeGUI or know of another way of adding a track to an MP4 file?

    Read the article

  • How to export or view audio file references in a PDF?

    - by redshift
    I have a an interactive PDF file that is over 90+ pages long. Each page is a map with city names that contains a Spanish pronunciation of that city in a .wav file. I'd say there are about 10-15 audio files for each map which comes out to 1000+ audio files. Is there a way to extract/export a list of the sound file names associated with each map? I tried to save the PDF to an HTML file, but it only exported images and text, and because the audio files were embedded in the PDF, the file names did not carry over to the HTML file. Any other ideas? I need to see what audio file goes with what map/page.

    Read the article

  • how to prevent MPMoviePlayer controls from hiding

    - by huevos de oro
    I am trying to implement a custom MPMoviePlayer to play mp3 audio. I have got it working in portrait mode along with an overlay window over the native controls - thanks to other stackoverflow posts. The current issue is the song progress control shows up when the media window opens (blue bar taking up the first 40 odd pixels), but seems to disappear when the song starts leaving a white bar. It will then re-appear when touching the area, so functionally works fine. I would like to find a way to ensure the controls always stay visible but have not found an appropriate property in the reference. Ideally I would like to have my custom control to replace the default, more because I would like to change the position that the look and feel. This being said, I understand it is not possible as the current position in the song from a MPMoviePlayer cannot be accessed.

    Read the article

  • Why is AudioOutputUnitStart freezing my app in iOS 4?

    - by Luke
    Hi guys, I have an audio app which uses the RemoteIO AudioUnit. It works fine on iPhone, iPad, and any flavor of the simulator on 3.2, but when it hits AudioOutputUnitStart (), it freezes. I get the message "AddRunningClient starting device on non-zero client count" in the console, which I'm not sure how to resolve. I stop the unit and dispose of the AudioComponent every time the app closes. The app works fine the first time I run after restarting everything, but freezes every time after that. What's strange is there are no error messages - just an unresponsive interface and a frozen line of code. Thanks for your help. Luke

    Read the article

  • Alternate play / pause button for WordPress wpaudio soundmanager plugin

    - by j-man86
    Hello! I am using the wpaudio plugin to convert mp3 links into a javascript/flash audio player. My problem is that I use this plugin in two areas on my site: one on a black background, and one on a white background. I need to use an alternate set of play/pause buttons for each page (white buttons for the black background and vice versa). I am at a total loss on how to do this. I need to some how incorporate a "if page is..." statement into the wpaudio.js but I don't know how to do this with jQuery. Can anyone help? Thanks so much!

    Read the article

  • Manipulating multi-track ogg files programatically

    - by Chad Birch
    I'm planning to create a program for manipulating multi-track OGG files, but I don't have any experience with the relevant libraries, so I'm looking for recommendations about which language/library to use for this. I don't really have any preference for the language, I'll happily code it in C, C#, Python, whatever makes things the easiest (or even possible). Perhaps it's even a possibility to automate Audacity somehow? In terms of requirements, I'm not looking for anything particularly fancy. It will probably be a command-line program, I don't need to be able to play the audio, draw image representations of the waveforms, etc. The program will basically be used as a converter, but I need to do some processing before outputting. That is, I need the ability to programatically remove some tracks, set panning per-track, change track volumes, etc. Nothing too complex, just some basic processing, and then output the result in either MP3 or a format easily converted to MP3, such as WAV. Any suggestions or general information would be appreciated, thanks.

    Read the article

  • AudioQueueOfflineRender returning empty data

    - by hyn
    I'm having problems using AudioQueueOfflineRender to decode AAC data. When I examine the buffer after the call, it is always filled with empty data. I made sure the input buffer is valid and packet descriptions are provided. I searched and found that a few others have had the same problem: http://lists.apple.com/archives/Coreaudio-api/2008/Jul/msg00119.html Also, the inTimestamp argument doesn't make sense to me. Why should the renderer care where in the audio the beginning of the buffer corresponds to? The function throws an error if I pass in NULL, so I pass in the timestamp anyway.

    Read the article

  • How to extract semi-precise frequencies from a WAV file using Fourier Transforms

    - by Seisatsu
    Let us say that I have a WAV file. In this file, is a series of sine tones at precise 1 second intervals. I want to use the FFTW library to extract these tones in sequence. Is this particularly hard to do? How would I go about this? Also, what is the bast way to write tones of this kind into a WAV file? I assume I would only need a simple audio library for the output. My language of choice is C

    Read the article

  • Why doesnt R.raw.'songname' not work on android devices?

    - by James Rattray
    I have some media (Audio tracks) on an app... With file path 'R.raw.test' I use some code to get it into a mediaplayer... MediaPlayer.create(Textbox.this, R.raw.fly); And it works PERFECTLY on the Android Emulator... (Plays track on click of button) Can someone explain why, when I put it on my Archos (5 IT) it doesnt work at all? -As soon as the button is clicked, it crashes... Do you have to do something to file paths or what? Please help... Thanks alot... James

    Read the article

  • Write wave files to memory in Java

    - by Cliff
    I'm trying to figure out why my servlet code creates wave files with improper headers. I use: AudioSystem.write( new AudioInputStream( new ByteArrayInputStream(memoryBytes), new AudioFormat(22000, 16, 1, true,false), memoryBytes.length ), AudioFileFormat.Type.WAVE, servletOutputStream ); taking a byte array from memory containing raw PCM samples and a servlet output stream that gets returned to the client. In the result I get a normal wave file but with zeros in the chunk size fields. Is the API broken? I would think that the size could be filled in using the size passed in the audio input stream. But now, after typing this out I'm thinking its not making this info available to the outer write() method on AudioSystem. It seems like the AudioSystem.write call needs a size parameter unless it is able to pull the size from the stream... which wouldn't work with an arbitrary sized stream. Does anyone know how to make this example work?

    Read the article

< Previous Page | 105 106 107 108 109 110 111 112 113 114 115 116  | Next Page >