Search Results

Search found 1918 results on 77 pages for 'matt klein'.

Page 25/77 | < Previous Page | 21 22 23 24 25 26 27 28 29 30 31 32  | Next Page >

  • Establishing a web page bookmarking process - looking for ideas to improve

    - by Matt
    Like many others, I have a process for bookmarking web pages to read later. My requirements for web page bookmarking are: Ability to bookmark pages must be available from all (within reason) platforms - PC/browser, mobile device, etc. Bookmarks must be centrally stored (implicit from #2) so that I can read the bookmarks from anywhere/any device Full text of web pages must be stored Bonus features would be: Bookmarks and page content should be full text searchable Maintain an archive indefinitely Distinguish between what's read vs. unread Bookmarked page content is cleaned up, e.g. ads eliminated, unnecessary html removed, pages better formatted for reading My current process (which addresses most of these requirements) is as follows: I set up a Gmail account with 2 labels, "Bookmarks Unread" and "Bookmarks Read" Gmail filters set up such that depending on the form of the address (using Gmail's '+string' functionality in addresses), the incoming bookmark gets labeled appropriately On each of my browsers/devices, I have an address book entry for [email protected] and [email protected]. If I want to clean up the page content, I use the Readability bookmarklet which does a great job of giving me the essential content only Anywhere I have Firefox, I use the Send Page by Email extension which, with 2 clicks, allows me to send the cleaned-up Readability page URL and content to one of the above email addresses. Where I don't have Firefox (e.g. iPhone or other mobile device) I use the native ability to send the current link via email (most/all apps have them, including the browser, RSS readers, NYTimes, etc.). In most cases (unless it's built into the particular app), this won't include the page body. The process is almost perfect. I've got the central access and ubiquitous access of Gmail as the storage mechanism, full text searchability (due to Gmail, but of course only for the URLs I send from that Firefox extension), a cleaned up page due to Readability, ability to read offline (assuming I use an IMAP client against Gmail) and permanent archiving of content, including what's been read vs. unread. The missing pieces are: The Send Page by Email Firefox extension seems to only send X bytes of a web page. Or some portion. So it limits my full text searchability. Where I don't have Firefox, I can only send the link, so no full text search at all in those cases. Instapaper looks like it meets most of my requirements (and bonus items). The only downside to me (personal preference) is that central storage is based on Instapaper vs. something more broad like Gmail, which as a generalized service and with Google behind it pretty much means it's permanent. I'm not too hung up on this, but I would definitely prefer to keep Gmail if possible. An upside of Instapaper is that it does the page clean-up as well as stores the entire page content, unlike my Firefox extension. Thoughts on addressing the gaps and improving this process further?

    Read the article

  • Instructions per cycle?

    - by Matt Simmons
    I've been learning a little bit more about how processors work, but I haven't been able to find a straight answer about instructions per cycle. For instance, I was under the impression that a four core CPU could execute four instructions per cycle, so a four core CPU running at 2Ghz would execute 8 billion operations per second. Is this the case? I'm sure it's oversimplifying things, but if there's a guide or something else I can use to set myself straight, I'm definitely open to ideas.

    Read the article

  • Is there a remote desktop management tool that I can email to people?

    - by Matt 'Trouble' Esse
    I often need to remotely manage PC and Macs for desktop support. I'm after a remote desktop management support tool that I could email (or send a url) that the customer could click on (or run) and I could then remotely manage their PC/Mac A tool that could work on both operating systems would be great but not mandatory (a separate tool for both/either will suffice) A tool that has an iPhone App would be fantastic too but this would just be very much a 'wish list' Looking forward to your suggestion!

    Read the article

  • Monitor displays at 1024x768; scrolls screen for higher resolutions

    - by Matt
    I have a dual monitor setup. Normally, they both display at 1680x1050. They have been setup this way for about a year. I'm using Windows XP Professional 2003 x64 SP2. Today, out of nowhere, one of the monitors kicked back to a lower resolution. I was not playing with any configuration at the time.. in fact all I had done was close a window (maybe a browser). But the thing is that the resolution is still preserved partially by the fact that the screen will scroll when you move the mouse. So it's like looking through a 1024x768 window into a 1680x1050 world. The monitor itself does not appear to be damaged, because I also have it connected to my netbook (via KVM) and higher resolutions work fine. I tried uninstalling/reinstalling the drivers to no avail. System restore doesn't help either. I'm unsure of the exact ATI card I'm using.. Device Manager lists it as "Radeon X300/X550/X1050". There is no Catalyst Control Center software installed. I tried to install it, but there doesn't seem to be a way to install it by itself ... it forces you to install another driver, which breaks both of my displays, forcing me to go into safe mode and run system restore again. Any ideas? Thanks

    Read the article

  • Migrate users from one Active Directory domain to another?

    - by Matt
    I work for a company that hosts desktops for a number of different companies. At the moment, all the clients access a single domain controller called HOSTING. Under that are groups for each company. Each of the hosting servers exist on the same network and so are therefore potentially browseable by other terminal servers. This has raised some security issues and I've found it a little tricky to manage the security. As well, it's possible to see who the other hosted companies are even though other users cannot see their data. What I'd like to do is isolate each clients terminal server/s into their own VLAN. In addition, I'm thinking that each TS would have it's own DC which could just run on the TS for that company. Overhead for a DC is fairly minimal. This would isolate users on that TS from seeing the other companies completely. Firstly, does this sound like a sensible plan? Second... if it is sensible, how would I go about pulling the accounts from the HOSTING domain to a new domain? ideally, without the need for users to change their passwords?

    Read the article

  • How to disable "safely remove hardware"

    - by Matt
    I have some windows 7 virtual machines in xen that have devices showing up in "safely remove hardware". I don't want users to ever be able to remove/eject any hardware at all. I'm told vmware has a hotplug option. xen doesn't seem to provide this for pci passthrough devices, therefore I'm looking for a reliable solution to prevent users from ejecting devices. This issue is not necessarily related just to virtual machines but seems to be a common problem with devices that get wrongly reported as removable. I'm ideally looking for a way to prevent all devices from appearing or just prevent the safely remove hardware option from ever coming up. I've tried setting device capabilities for specific devices on boot with a script but this for some reason doesn't always seem to work reliably. Is there a way to prevent this icon from appearing in the notification area completely, either by registry key or group policy? I should point out that setting this in group policy to "Administrators" did not seem to work. [Computer ConfigurationWindows SettingsSecurity SettingsLocal PoliciesSecurity Optionsevices:Allowed to format and eject removable media]

    Read the article

  • How to specify the wget file location (output location)

    - by Matt
    i have this batch file that reads an input file to grab images from a URL. How can I specify a different output location for the wget files. @echo off setlocal enabledelayedexpansion for /f %%l in (Input.txt) do ( set line=%%l wget -O !line:~0,10!.jpg !line:~10! ) I want the output to be in a different location. edit If wget is run from C:\wget\bin. How can the output be saved to C:\folder\folder\x

    Read the article

  • dual/multi-boot computers and software licensing

    - by Matt
    Suppose you have a computer with two or more operating systems, and a certain piece of software whose license terms allows it to be installed on one computer, and it does a daily check with a remote server to verify that your serial is only used on the original install computer. You install this software on each of your OSes, but since its a different OS the remote server would have to determine that it is not on the same computer, and so would disable your license. So my question, when a license refers to a single computer, does a situation like this usually count as a single computer, or do the multiple OSes sort of make it multiple computers? How do you think a software vendor (specifically thinking AV companies that do this sort of serial check) would handle this situation?

    Read the article

  • If I use a video format converter to change a movie from AVI to MKV, will quality stay the same?

    - by Matt
    I know they're both container formats and what matters is the actual codec used, but what I don't know is if video converting software will do anything to change the codec, or if it just repackages. The reason I need to know is that I have several .avi files with subtitle files, and I'm wanting to turn them into .mkv so I can attach the subs and not need a second subtitle file anymore. Will my new .mkv files be identical in video and audio quality to the original .avi?

    Read the article

  • Web application and remote storage of files

    - by Matt
    Hi have a web application that can store lots and lots of files on the server. i.e. users upload data to it. The files are stored below a particular storage path. The web host will be an IBM xseries 345. However, the disks are really expensive so we would like to put the files onto a less expensive server. Now here is the question. Should I use an NFS mount on the IBM server of a path on the storage server? Or should I write some scripts to upload the files to the storage server instead. Both the storage server and the web host are on the same network. Only the web server is visible to the world. Is NFS performance suitable for an expected low to moderately loaded server?

    Read the article

  • Traffic shaping & monthly traffic limit in Tomato?

    - by Matt H
    Is there a way to do a monthly traffic limit in Tomato, DDWRT or OpenWRT in addition to the regular QoS? This is for a house with several students sharing the internet. I.e. for a specific IP address, IP Range or MAC address, the firmware will count the download traffic for that month. When a configurable limit is set, it'll either limit it to say 64kbit/s up/down or drop all traffic and maybe redirect web traffic to an internal web server telling them that they have exceeded their quota. How can this be done with those firmwares?

    Read the article

  • How do i enable innodb on ubuntu server 10.04

    - by Matt
    Here is my entire my.cnf [client] port = 3306 socket = /var/run/mysqld/mysqld.sock # Here is entries for some specific programs # The following values assume you have at least 32M ram # This was formally known as [safe_mysqld]. Both versions are currently parsed. [mysqld_safe] socket = /var/run/mysqld/mysqld.sock nice = 0 [mysqld] key_buffer = 224M sort_buffer_size = 4M read_buffer_size = 4M read_rnd_buffer_size = 4M myisam_sort_buffer_size = 12M query_cache_size = 44M # # * Basic Settings # # # * IMPORTANT # If you make changes to these settings and your system uses apparmor, you may # also need to also adjust /etc/apparmor.d/usr.sbin.mysqld. # user = mysql socket = /var/run/mysqld/mysqld.sock port = 3306 basedir = /usr datadir = /var/lib/mysql tmpdir = /tmp skip-external-locking # # Instead of skip-networking the default is now to listen only on # localhost which is more compatible and is not less secure. bind-address = 127.0.0.1 # # * Fine Tuning # #key_buffer = 16M max_allowed_packet = 16M thread_stack = 192K thread_cache_size = 8 # This replaces the startup script and checks MyISAM tables if needed # the first time they are touched myisam-recover = BACKUP #max_connections = 100 #table_cache = 64 #thread_concurrency = 10 # # * Query Cache Configuration # query_cache_limit = 1M #query_cache_size = 16M # # * Logging and Replication # # Both location gets rotated by the cronjob. # Be aware that this log type is a performance killer. # As of 5.1 you can enable the log at runtime! #general_log_file = /var/log/mysql/mysql.log #general_log = 1 log_error = /var/log/mysql/error.log # Here you can see queries with especially long duration #log_slow_queries = /var/log/mysql/mysql-slow.log #long_query_time = 2 #log-queries-not-using-indexes # # The following can be used as easy to replay backup logs or for replication. # note: if you are setting up a replication slave, see README.Debian about # other settings you may need to change. #server-id = 1 #log_bin = /var/log/mysql/mysql-bin.log expire_logs_days = 10 max_binlog_size = 100M #binlog_do_db = include_database_name #binlog_ignore_db = include_database_name # # * InnoDB # # InnoDB is enabled by default with a 10MB datafile in /var/lib/mysql/. # Read the manual for more InnoDB related options. There are many! # # * Security Features # # Read the manual, too, if you want chroot! # chroot = /var/lib/mysql/ # # For generating SSL certificates I recommend the OpenSSL GUI "tinyca". # # ssl-ca=/etc/mysql/cacert.pem # ssl-cert=/etc/mysql/server-cert.pem # ssl-key=/etc/mysql/server-key.pem [mysqldump] quick quote-names max_allowed_packet = 16M [mysql] #no-auto-rehash # faster start of mysql but no tab completition [isamchk] key_buffer = 16M # # * IMPORTANT: Additional settings that can override those from this file! # The files must end with '.cnf', otherwise they'll be ignored. # !includedir /etc/mysql/conf.d/ And here is my show engines call....i have no idea what i need to do to enable innodb show engines; +------------+---------+----------------------------------------------------------------+--------------+------+------------+ | Engine | Support | Comment | Transactions | XA | Savepoints | +------------+---------+----------------------------------------------------------------+--------------+------+------------+ | MyISAM | DEFAULT | Default engine as of MySQL 3.23 with great performance | NO | NO | NO | | MRG_MYISAM | YES | Collection of identical MyISAM tables | NO | NO | NO | | BLACKHOLE | YES | /dev/null storage engine (anything you write to it disappears) | NO | NO | NO | | CSV | YES | CSV storage engine | NO | NO | NO | | MEMORY | YES | Hash based, stored in memory, useful for temporary tables | NO | NO | NO | | FEDERATED | NO | Federated MySQL storage engine | NULL | NULL | NULL | | ARCHIVE | YES | Archive storage engine | NO | NO | NO | +------------+---------+----------------------------------------------------------------+--------------+------+------------+ 7 rows in set (0.00 sec)

    Read the article

  • Should I anticipate any problems trying to use the same SSL Cert on 2 computers (primary, backup)?

    - by Matt
    We have a production machine running IIS6 with a wildcard SSL certificate. The certificate that was installed is not exportable. We want to upgrade the system to IIS7. As part of this venture, we're creating a backup/failover server that will serve the exact same websites - when we take the primary down for upgrade, the secondary will take over. As such, the secondary also needs the SSL certificate. However, since the certificate was not exportable, this means re-keying it from Go Daddy. Per http://help.godaddy.com/article/867, I know that by re-keying the certificate the original will stop working. I'm still pretty new to SSL certificates, so are there any problems I should anticipate when installing the same SSL certificate on 2 different machines?

    Read the article

  • non-interactively upload file to sftp server, using password

    - by matt
    Hello Guys, I know, this is not the recommended way to do this. But, I do not have another choice: I've got to set up a cron job that will regularly upload a file to an external SFTP Server (no FTP available, and I do only have a username/password for it but no key.) Still, I need to set up a cron that will regularly connect to that sftp and upload a file. sftp <<EOF put filename exit EOF therefore will not work, because sftp asks for the password, before STDIN is evaluated. What can I do, to pass the Password to sftp? Again: I am aware of the potential security risk, but I really do not have any choice here, and the server from which the file is uploaded is protected rather well.

    Read the article

  • Apache mpm-itk Performance

    - by Matt Beckman
    I manage a bunch of VPSs with memory ranging from 1GB to 8GB. Most of these websites are Joomla websites, and the servers must support multiple sites/users/S-FTP. I use mpm-itk almost exclusively (mostly due to it's convenience in these shared environments). However, I'm aware it isn't known for performance, so I need some advice on making it faster. Due to the lack of documentation when I first went the way of mpm-itk, I included only one setting in the config, and that was to limit each user to 50 clients (the rest I left up to defaults): <IfModule mpm_itk_module> MaxClientsVHost 50 </IfModule> Are there any better alternatives available? Are there any settings supported in mpm-prefork or mpm-worker that are also supported in mpm-itk? Thanks!

    Read the article

  • How to get Hyper-V to recognize that a VM isn't stopping, it's stopped.

    - by Matt Burke
    I have a Hyper-V virtual machine that has some problems. First, its settings are apparently corrupt: The WMI object contained an invalid value in property BIOSNumLock. The WMI object contained an invalid value in property AutomaticStartupAction. Second, I saw something that implied that I could toggle the NumLock checkbox on the BIOS property page to fix the first problem, but the checkbox didn't show up. So I connected to the VM and shut it down. Now (many hours later) Hyper-V claims that the VM is "Stopping". However, when I follow the interweb's directions on forcing a VM offline (i.e. find vmwp and kill it), I can't find vmwp. The VM's WMI object has a PID that doesn't exist, and none of the vmwp processes have this machine's GUID. So, how do I trick Hyper-V into seeing that the machine is offline (or letting me delete it)?

    Read the article

  • How can Rackspace beat DigitalOcean's Pricepoint? [on hold]

    - by Matt Jensen
    I have recently discovered Digital Ocean and I have found it to be a relatively nice experience for small staging servers, and the thought occurred to me, why am I paying $267~ for a server on Rackspace (40GB RAM, 160 GB Drive, 2 vCPUs, 400 Mb/s) when Digital Ocean offers a server for $40 (40 GB RAM, 60GB Drive [storage is not a concern of mine], 2 vCPUs, ?Mb/s)? Does Rackspace offer some kind of obvious advantages in Transfer speed/bandwidth? My applications are small startups that for the immediate future will only have about 200-300 concurrent users at once.

    Read the article

  • Is there any way to disable certain keyboard shortcuts in Windows 7?

    - by matt b
    There is something going on with my Windows 7 machine where attempting to unlock the workstation the Win key is "stuck" on (not quite sure what is causing this). What this causes is that when I am entering my password, which contains a "p", Windows thinks I am pressing Win + p as if I want to change my display settings to enable a projector. This is highly annoying when attempting to unlock my workstation! The workaround I've found to "unstick" the Win key is to cycle through the projector/display settings once, land back on the original value, and then resume entering my password as normal. But is there any way to completely disable the shortcut for Win + p ? In all likelihood I will never hook up a projector to this machine, or on the once-in-a-lifetime chance that I do, I am more than happy with having to go into the Display Settings myself.

    Read the article

  • Can't port forward ssh on Asus RT-NU56 wireless router

    - by Matt
    I cannot ssh in (using putty) when I use 10.0.1.31 as the address. database server - wired switch part of Asus wireless router - office LAN So basically, we want to ssh in to the database server from our computer on the wired office LAN. Asus router has an ip of 10.0.1.31 and database server has a static IP of 192.168.0.20 I set up port forwarding like this: ssh 22 192.168.0.20 22 BOTH Firewall is turned completely off. Any other settings I'm missing?

    Read the article

  • How do I mklink junction + move content from C:\Program Files to D:\Program Files?

    - by Matt
    I have a few applications that absolutely refuse to install into anything but C:\Program Files or C:\Program Files(x86). Changing the registry keys for default install folders doesn't seem to provide any satisfaction and so now I'm wondering about throwing a NTFS junction in there to force these pesky applications to cooperate. There are files currently in use within Windows so it's quite likely I am not going to be able to do this within the active OS. Is there some bootable Windows 7 system tools that would allow me to make this happen? Seems I will need the ability to copy files (with permissions!) from one drive to another, as well as make the junction for Windows.

    Read the article

  • Can't double click files to open them in inDesign (CS5)

    - by Matt
    I cannot open a file unless I open inDesign (the program) and then do File-Open If I double click, it starts to open, then just hangs forever. AFTER I close it, and look in the directory where they're saved, I see a (temporary?) "lock" file. Now I can double click the original file and it opens just fine. However, now when I close iD it deletes the file and the whole process starts again... I have tried updating the software, uninstalled COMPLETELY and reinstalled, tried a brand new Win7 install. These files are all saved on a network drive, the computer is a new quad-core Dell with 12GB of RAM and a fresh x64 Win7 install on the SSD. Does not happen with other programs.

    Read the article

  • VLC RTP Streaming in FC12

    - by Matt D
    I'm trying to get VLC to work streaming RTP audio/video over my office network. The goal is multicast a/v streaming. In all test cases, we are streaming from VLC to VLC. I am able to stream from Windows to Windows, and from Fedora to Windows, but not from Windows to Fedora. Additionally, I am unable to receive a LOCAL stream from one instance of VLC to another, within Fedora. I don't see any reason why this would be. The buffer indicator (where the elapsed/total time is normally displayed) never shows any connectivity, so it would appear to be a network problem, but since I am able to stream from Fedora to Windows (same IP, same port) I thought it would be something else. Does anyone know of a solution to this issue?

    Read the article

  • Is it possible to manually install RDP client 6.1 for platforms that aren't directly supported?

    - by Matt
    I have some clients that need RDP client 6.1 in order to utilize the new easy print driver. However, the installer doesn't allow it to install on several platforms such as Windows Home Server or Windows Server 2003 because they are not XP. The version check prevents it going further. I'm reasonably confident that it should actually run however and want to try it. Has anyone done this before? what I should really ask is... what files should I manually copy (backing up originals first)... just the exe or are there lots of dependant dll's that need upgrading too?

    Read the article

< Previous Page | 21 22 23 24 25 26 27 28 29 30 31 32  | Next Page >