Search Results

Search found 11197 results on 448 pages for 'handle leak'.

Page 137/448 | < Previous Page | 133 134 135 136 137 138 139 140 141 142 143 144  | Next Page >

  • Run Java Project from Ubuntu Terminal?

    - by Christopher Gwilliams
    I have a small java project that handle connections. In order to run it from the terminal I have to cd into the folder that contains the source and run the following command: java -cp classes com.packagename.mainclass Where classes is the folder that contains the classes. I want ubuntu to run this application on startup, is there a Java command I can use? Or am I just better off creating a shell script? Thanks!

    Read the article

  • Tracking history in excel.

    - by Vojtech R.
    Hi, there is any possibility to track changes in Excel XP? With time, content and name of editor? Excel is running under just one windows account so system of revisions cant handle name of editors. Thanks

    Read the article

  • Increase Audio volume on Avi files.

    - by Arch Angel
    Is there a simple way to adjust audio volume in avi files, like using MP3gain so i can set them all the same it is a problem having to keep adjusting the volume. I have tried VirtualDub but it doesnt handle my avi files as they are encoded in XVID

    Read the article

  • Port 5357 TCP on Windows 7 professional 64 bit?

    - by Registered
    Is there a reason this port is open, a quick Nmap scan and Nessus scan reveal it's open, why? Are there any ramifications if I close this port via the firewall rule set? Or does anyone here now more info about this port besides Google? WTF? 1)http://www.symantec.com/connect/blogs/who-left-tunnel-door-open-windows-firewall-vista-0 I know the talk is about Vista, but I am pretty sure it's the same port on 7, also. 2)Port 5357 common errors:The port is vulnerable to info leak problems allowing it to be accessed remotely by malicious authors. (Web Services for Devices) I am blocking this crap, if I have issues will just re-enable. Damn windows. Inbound rule for Network Discovery to allow WSDAPI Events via Function Discovery. [TCP 5357] You just got blocked, until I break something, will see. Time to re-Nmap and re-Nessus. Nmap scan 0 open ports after closing Port 5357,Win7 still works for now, one more scan with Nessus just to make sure all is well.

    Read the article

  • Run a local script on a remote server using ssh with out having to worry about quotes

    - by Michael Irey
    So I have been running local scripts fine on a remote server: ssh user@server '`cat local-script.sh`' However, today I have a script that has both single and double quotes in it. Which causes the script to fail because the output of cat local-script.sh is wrapped in quotes. With out modifying the script itself, is there a better way to handle this? I thought this may work: ssh user@server $(<local-script.sh) But is does not seem to do anything...

    Read the article

  • Java: very slow tomcat and too big war file

    - by NaN
    I created some sort of RESTful API backend for a mobile app. It's written completely in Java using Jersey as Framework. At the moment no database is used, it's all in the memory, but this is no problem so far (it's only for prototyping purposes). I ordered the smallest package from digital ocean and installed tomcat7. All in all tomcat works, but I have three major problems: 1) It takes a long time until tomcat deploys the app: I deploy it per tomcat manager and it takes about 2 minutes unit the site works (excl. war upload time). 2) The war files are quite big (16MB): I don't know why they are so big. There are no database dependencies and most logic is written in plain java. Okay, we are using jersey, but 16MB are a lot for the logic of a small webservice. 3) I have to restart tomcat all 3 days or so. It looks like a memory leak or something similar. If the app runs for a few days the response time is quite high and the server seems to be frozen. It works again, if I restart tomcat per ssh. You can find my mvn pom file right here. Do you have some tips? Are there good tomcat alternatives?

    Read the article

  • managing print account for ubuntu server

    - by JGC
    hi I want to setup a system for a domain which enable me to charge money into user's account for printing. I am using Ubuntu server 8.04 and installed samba server for creating domain but I don't know how to handle my problem. does anyone has any idea?

    Read the article

  • Open Source MS Word Password Recovery?

    - by squillman
    Does anyone know of a free open source recovery tool? Should include recovery of document protection password ideally. EDIT: I'll settle for just Free instead of free and open source :) I'm looking for something that will handle not only passwords used to open the file but also passwords used to do Word's different forms of document protection.

    Read the article

  • What Are All the Variables Necessary to Create Blackbox Logs for Nginx?

    - by Alan Gutierrez
    There's an article out there, Profiling LAMP Applications with Apache's Blackbox Logs, that describes how to create a log that records a lot of detailed information missing in the common and combined log formats. This information is supposed to help you resolve performance issues. As the author notes "While the common log-file format (and the combined format) are great for hit tracking, they aren't suitable for getting hardcore performance data." The article describes a "blackbox" log format, like a blackbox flight recorder on an aircraft, that gathers information used to profile server performance, missing from the hit tracking log formats: Keep alive status, remote port, child processes, bytes sent, etc. LogFormat "%a/%S %X %t \"%r\" %s/%>s %{pid}P/%{tid}P %T/%D %I/%O/%B" blackbox I'm trying to recreate as much of the format for Nginx, and would like help filling in the blanks. Here's what Nginx blackbox format would look like, the unmapped Apache directives have question marks after their names. access_log blackbox '$remote_addr/$remote_port X? [$time_local] "$request"' 's?/$status $pid/0 T?/D? I?/O?/B?' Here's a table of the variables I've been able to map from the Nginx documentation. %a = $remote_addr - The IP address of the remote client. %S = $remote_port - The port of the remote client. %X = ? - Keep alive status. %t = $time_local - The start time of the request. %r = $request - The first line of request containing method verb, path and protocol. %s = ? - Status before any redirections. %>s = $status - Status after any redirections. %{pid}P = $pid - The process id. %{tid}P = N/A - The thread id, which is non-applicable to Nignx. %T = ? - The time in seconds to handle the request. %D = ? - The time in milliseconds to handle the request. %I = ? - The count of bytes received including headers. %O = ? - The count of bytes sent including headers. %B = ? - The count of bytes sent excluding headers, but with a 0 for none instead of '-'. Looking for help filling in the missing variables, or confirmation that the missing variables are in fact, unavailable in Nginx.

    Read the article

  • debian LTS : how to keep packages up-to-date with last security fixes using apt

    - by Quentin
    What's the best way to keep packages up-to-date (ie with the last security fixes) without worried about major version update ? For instance, apache2 for squeezeis is 2.2.16 (https://packages.debian.org/source/squeeze/apache2) However, last apache2 version for the 2.2.x branch is 2.2.27 Test repository can't be used since they use the 2.4.x versions and I'd like to stick on the 2.2.x (to avoid migration issues) How would you handle this situation and how can I update to 2.2.27 ?

    Read the article

  • Virtual Box between differnt PCs

    - by Jim
    I am using Virtual Box machine (as .vdi) on external HDD drive so I can move it between different pcs. On each PC I have created new VM machine, attached that vdi and it seems working well. However, I've noticed that during my last transfer the latest installed applications were not saved on the .vdi, which made me to loose 1 day of work. I don't remember of doing any snapshots which I could merge. Can someone give me a hint how to handle this efficiently in feature?

    Read the article

  • dynamically allocating space for variable in ksh

    - by Govind Rajan
    I am new to ksh unix shell scripting.I need my ksh script to get inputs from the user. and the number of inputs is unkown. I need to get the inputs and process again further .The sample run could be : Enter the input values : Hello how are you Select two values from the above : how you Here the italized text are the inputs from the user.The inputs number given by the user is unknown .I down know how to handle the values given by user. Thanks in advance

    Read the article

  • iMac memory limit

    - by Mike
    I have an iMac that was from the first generation of aluminum iMacs. The reported model is "iMac 7,1". This iMac's manual says I can put 2 2GB modules, but when this manual was made we don't have modules with more than 2GB and also we had Leopard then, that I suppose can handle less memory than Snow leopard. Today we have 4GB modules, so can I put two 4GB modules and make it 8GB? thanks.

    Read the article

  • Is ThinkPad X61T SD card slot SDXC compatible?

    - by trismarck
    I have a Thinkpad X61T 7763 tablet and I was planning to buy a 64GB card for the SD card slot. The 64GB card supports the new SDXC standard and I'm not sure if the SD card driver / hardware can handle that. So far, I've successfully used 8GB cards in the slot, but they were SDHC ones. Thinkpad manuals don't give the answer. I wonder if anyone tried this configuration already and what was the outcome.

    Read the article

  • MAX Connection Pool Setting SQL Server 2008

    - by dkeeshin
    We are expecting a large number of users to hit a Website built with IIS/.Net 4.0 that our SQL Server 2008 database server is providing data for. The database is around 2GB in size. We are contemplating increasing the MAX CONNECTION POOL to between 500 to 1000 -- to handle the estimated traffic. Two questions (1) Does anyone have any hard performance numbers indicating the kind of improvement this may provide? (2) What is the impact of hitting that MAX CONNECTION POOL number in a production enviroment?

    Read the article

  • Multiple OS's and GRUB chainloading

    - by Kent
    Hi, I want to have multiple OS installations and I have been advised that chain loading using GRUB is a good way to handle this. I have looked at tutorials on the web but I still have some questions before I can start. I want: Windows XP: 20 GB. For running some school stuff and a game which does not work through WINE. Xubuntu 9.04: 85 GB. My main OS. Another Linux distribution: 15 GB . For experimenting and trying Linux distributions out. I will: Wipe and install various distributions quite often on the 15 Use dd to make a copy of my Windows partition after installing it and getting things to work as I like. My experience is that Windows needs to be re-installed maybe once per year to not get bloated and slow. I have been told: To use GRUB chain loading. It will make it easier when kernel upgrades are made in the Linux distributions, as they modify the GRUB boot-menu. To my understanding I need to: (I might very well be mistaken) Install Windows first. Then install Xubuntu and let it write over the MBR with GRUB (I guess this is the default). Get the GRUB on the MBR start Windows XP if I want to (it's done by default), start Xubuntu using the kernel of my choice or defer execution to the boot sector of my other Linux distribution. The actual chain loading will only occur when I want to start my experimental install of Linux. I wonder: Is step 3 above correct and a good way to handle this? Is it also a good way to use chain-loading for both Xubuntu and my experimental Linux installation? How do I get a Linux distribution to install the boot loader it comes with to the boot sector of its partition and not to the MBR? If I can't get it to not touch the MBR. Then I could make a backup of the MBR using dd and then write it back after installing my experimental Linux installation. But then, how would I get the boot loader (lets say GRUB) into the boot sector of the experimental Linux installation? How would it work if said Linux installation gets a new kernel update and needs to update the GRUB menu?

    Read the article

  • FTP passive mode with limited port range?

    - by Phil Keeling
    I am running FileZilla FTP Server with passive mode enabled, and due to firewall constraints I have to limit the passive mode port range to only 6 ports. My question is how would FileZilla handle any situation where more than 6 concurrent FTP connections are active and want to passively upload a file. Would it queue the connections and prioritise them in a first in, first out manner? I'm not too familiar with FTP so any insight would be appreciated.

    Read the article

  • How Can We Create Blackbox Logs for Nginx?

    - by Alan Gutierrez
    There's an article out there, Profiling LAMP Applications with Apache's Blackbox Logs, that describes how to create a log that records a lot of detailed information missing in the common and combined log formats. This information is supposed to help you resolve performance issues. As the author notes "While the common log-file format (and the combined format) are great for hit tracking, they aren't suitable for getting hardcore performance data." The article describes a "blackbox" log format, like a blackbox flight recorder on an aircraft, that gathers information used to profile server performance, missing from the hit tracking log formats: Keep alive status, remote port, child processes, bytes sent, etc. LogFormat "%a/%S %X %t \"%r\" %s/%>s %{pid}P/%{tid}P %T/%D %I/%O/%B" blackbox I'm trying to recreate as much of the format for Nginx, and would like help filling in the blanks. Here's what Nginx blackbox format would look like, the unmapped Apache directives have question marks after their names. access_log blackbox '$remote_addr/$remote_port X? [$time_local] "$request"' 's?/$status $pid/0 T?/D? I?/$bytes_sent/$body_bytes_sent' Here's a table of the variables I've been able to map from the Nginx documentation. %a = $remote_addr - The IP address of the remote client. %S = $remote_port - The port of the remote client. %X = ? - Keep alive status. %t = $time_local - The start time of the request. %r = $request - The first line of request containing method verb, path and protocol. %s = ? - Status before any redirections. %>s = $status - Status after any redirections. %{pid}P = $pid - The process id. %{tid}P = N/A - The thread id, which is non-applicable to Nignx. %T = ? - The time in seconds to handle the request. %D = $request_time - The time in milliseconds to handle the request. %I = ? - The count of bytes received including headers. %O = $bytes_sent - The count of bytes sent including headers. %B = $body_bytes_sent - The count of bytes sent excluding headers, but with a 0 for none instead of '-'. Looking for help filling in the missing variables, or confirmation that the missing variables are in fact, unavailable in Nginx.

    Read the article

< Previous Page | 133 134 135 136 137 138 139 140 141 142 143 144  | Next Page >