Search Results

Search found 13293 results on 532 pages for 'small ticket'.

Page 254/532 | < Previous Page | 250 251 252 253 254 255 256 257 258 259 260 261  | Next Page >

  • Moving to the next line to populate an excel file from VBA

    - by edmon
    I have the below code that takes certain fields from my MS Access (A small Hotel Reservation Database)form and populates defined cells in the said Excel file. Dim objXLApp As Object Dim objXLBook As Object Set objXLApp = CreateObject("Excel.Application") Set objXLBook = objXLApp.Workbooks.Open("Y:\123files\File\Hotel Reservation.xls") objXLApp.Application.Visible = True objXLBook.ActiveSheet.Range("B2") = Me.GuestFirstName & " " & GuestLastName objXLBook.ActiveSheet.Range("C2") = Me.PhoneNumber objXLBook.ActiveSheet.Range("E2") = Me.cboCheckInDate objXLBook.ActiveSheet.Range("F2") = Me.cboCheckOutDate objXLBook.ActiveSheet.Range("H2") = Me.RoomType objXLBook.ActiveSheet.Range("I2") = Me.RoomNumber End Sub How can I keep populating a new Guest to the same Excel file just on the next row?

    Read the article

  • Good bitmap fonts with big sizes and unicode support

    - by bitonic
    I really like bitmap fonts for programming/terminal. As far as I know there are two bitmap fonts with good unicode support: Unifont Fixed The problem is that I have a really high resolution screen, and they're both too small. Fixed does include a large size (10x20) but it looks really bad (it's basically always bold, and bold is a different face). Are there any other bitmap fonts with unicode support and large sizes? Terminus is the only font with a decent size but it doesn't have good unicode support. Having good coverage for mathematical symbols would be enough, since that's what I need.

    Read the article

  • Time of splitting and encoding video with ffmpeg increases exponentially

    - by Rnd_d
    I'm trying to split video by subtitles and encode into .mp4(h.264/aac) using ffmpeg, but it takes so much time! First pieces of video are splited and encoded really fast, but for each iteration time increases, and so 40 min video takes all night or more. Small 3 min video takes max 10 mins. cmd for splitting and encoding: ffmpeg -i filename.avi -ss 00:00:0(time of sub start) -t 0:0:3(time of sub duration) -acodec libfaac -vcodec libx264 -bf 0 -f mp4 filename.mp4 ffmpeg version N-34849-g07c7ffc (last, i think) How I can make it faster? Are there, maybe, some magic arguments for ffmpeg or some hacks?

    Read the article

  • Ubuntu+Mono+Postgres+ASP.NET 4.0. No problem?

    - by wreck_of_u
    Would this be ok? I'm an ASP.NET developer and I'm planning to build "portable" web app servers based on Atom D510 mini-ITX. I have ran Ubuntu 10 with MySQL along with a separate IIS machines (win 2k3, 2k8) before with no problems. But now I'm thinking of "packaging" a web/db server into one small, cheap machine. I thought of Ubuntu/Mono/Postgres/ASP.NET, that it would be a good idea but I'm not sure? I have not actually tried it yet. Your thoughts?

    Read the article

  • How to collect figures of traffic used per-host, broken up by time and destination?

    - by Seishun
    We have a relatively small network, all PSs in one subnet. One PC with two NICs and pfSense installed works as a firewall/router. There is an OpenVPN tunnel to a remote location, created as a site-to-site connection to another pfSense box there. I have an assignment to capture, store and show (via a web interface) information on traffic generated (both incoming and outcoming) by each host on our subnet and present it in several views: megabytes per calendar hours / days / months / years (that is, not just "one month back", but "in Dec 2010" and so on); megabytes per destination: VPN to remote location / other destinations / Google Apps servers. I tried the software packages in pfSense that offer traffic montoring - but it seems they don't store the information fixed by months, instead showing the amounts of traffic generated in periods (days/months/etc) calculated from the present moment. I'm also interested in understanding what would be the best way for me to break up traffic by hosts and destinations. I'm open to all suggestions, even if they mean that I will have to understand something new to me.

    Read the article

  • AD Local Admins without password sharing

    - by Cocoabean
    My team is building out an Active Directory environment in a small grad school with support for general computer labs, and staff/faculty machine and account management. We have a team of student consultants that are hired to do general help desk work. As of now we have a local admin account on every machine. It has the same password and all of us know it. I know it's not best practice and I want to avoid this with the new setup. We want to have local admin accounts in case there are network issues that prevent AD authentication, but we do not want this account to be generic with a shared password. Is there a way we can get each machine to cache the necessary information to authenticate a group of local admins so that if AD is somehow inaccessible, student consultants can still login with their AD admin accounts?

    Read the article

  • OpenOffice - alternative paragraphs (two PDFs, one source)

    - by RedGrittyBrick
    I use OpenOffice to produce a document in PDF format. This document is occasionally revised and reissued. I now need to produce a variant of this document with a small number of paragraphs replaced by other paragraphs. I would prefer to avoid maintaining two separate documents where 95% of the text is the same. In other words, from one ODT file I would like to be able to produce two PDFs with slighly differing content. Are there some OpenOffice features I could use that would help with this?

    Read the article

  • Backup software for Ubuntu - which one?

    - by Industrial
    Hi everybody, I have spent some time testing out different backup solutions for my small home office during the last weeks, but still haven't found anything that have been working out too well yet. We can definitely work with a non-GUI script if that's what it takes, if only the requirements are fulfilled: Upload to Amazon S3 Europe. We get unbelievable slow uploading speed to US, so uploading 400+ GB of data will not be happening anytime this year... Incremental backups - only changed files shall be uploaded or we will have a big bill from Amazon in the end of each month.. Files should not be uploaded in one big per-folder archive. This is not efficient at all, since if we change one file in a subfolder, a huge two-digit GB sized file would have to be uploaded during next backup. Not good for economy again, or traffic overhead on our internet connection. What options are available to us? Thanks!

    Read the article

  • Bacula configuration for clients that are turned on and off randomly

    - by Rastloser
    I'm evaluating Bacula as a centralized backup tool for a small network where users will turn machines on and off unpredictably. Some of the headless Linux boxes I need to back up are intended to be turned off by pressing the on/off-button on the case, without any way of telling the user to wait for a backup job to finish. So, we don't know when backup jobs may run (anacron might help with this, right?) and we don't know whether they'll be allowed to finish. Is Bacula a reasonable choice for such an environment?

    Read the article

  • Set up router to vpn into proxy server

    - by NKimber
    I have a small network with a single LinkSys router connected to broadband in US via Comcast. I have a VPN proxy server account that I can use with a standard Windows connection, allowing me to have a geographic IP fingerprint in Europe, this is useful for a number of purposes. I want to setup a 2nd router that automatically connects via VPN to this proxy service, so any hardware that is connected to router 2 looks as though it is originating network requests in Europe, and any hardware connected to my main router has normal Comcast traffic (all requests are originating from USA). My 2nd router is a LinkSys WRT54G2, I'm having trouble getting this configured. Question, is what I'm trying to do even feasible? Should the WRT54G2 be able to do this with native functionality? Would flashing it with DD-WRT allow me to achieve my objectives?

    Read the article

  • Network Table assistance

    - by mitchnufc
    I am designing a small network and have came up with the following table I am just wondering if this seems right, would appreciate some feedback, thanks. Network/Router First IP Last IP Subnet Host Broadcast Router 1 162.10.0.1 162.10.0.7 255.255.255.248 162.10.0.0 162.10.0.8 Network 1 162.10.1.1 162.10.2.253 255.255.254.0 162.10.1.0 162.10.2.254 Network 2 162.10.0.9 162.10.0.14 255.255.255.248 162.10.0.8 162.10.0.15 Router 2 162.10.0.17 162.10.0.18 255.255.255.252 162.10.0.16 162.10.0.19 Network 3 162.10.0.21 162.10.0.146 255.255.255.128 162.10.0.20 162.10.0.147 Router one is the IP assigned by the ISP

    Read the article

  • How to serve a .php file locally?

    - by isomorphismes
    This part of the PHP documentation says that I should be able to make a small, fake server to serve up some local .php files in a folder using php -S localhost:8000 . But when I try that I get the following error: Usage: php [options] [-f] <file> [--] [args...] php [options] -r <code> [--] [args...] php [options] [-B <begin_code>] -R <code> [-E <end_code>] [--] [args...] php [options] [-B <begin_code>] -F <file> [-E <end_code>] [--] [args...] php [options] -- [args...] php [options] -a What am I doing wrong?

    Read the article

  • WSS 3.0 fails to hide quick launch items for which the current user does not have access

    - by Nils
    I'm running a Small Business Server 2008 with Windows Sharepoint Services 3.0 (WSS 3.0). I thought WSS was supposed to hide menu items for which the current logged in user don't have access? Apparently, all users can see all links, regardless of whether they have access. This applies to both links to newly created sub-sites as well as document libraries/lists. Is this expected behaviour, or is there a misconfiguration somewhere that causes the links to stay visible even for users without access? Thanks!

    Read the article

  • Asking for brief explanation of reverse proxying and recommended software

    - by 80skeys
    I need to set up a reverse proxy where the backend web server is serving https pages only. I've never set up a reverse proxy and would like a brief overview of how it works. One of my questions is whether the proxy needs to run https also, or does simple http suffice? Second question would be whether to use Apache, varnish, nginx, or squid. This is for an internal site for a small company, so not a lot of traffic expected. Maybe a few dozen users each day.

    Read the article

  • (Possibly) Corrupted DDCrypt self-decrypting file

    - by sca
    I recently discovered some files I had encrypted and archived on a CD using DDCrypt 2.0, which is a small encryption utility from back in the day. One of its functions is to create self-decrypting files, which are essentially EXE files that you can enter a password and then have the file extracted onto your filesystem. Unfortunately, I can't seem to get them to do this properly. They appear to be decrypting and then at the last moment I get the error: Error opening encrypted file C:\users\username\Temp\dde3cfa.tmp Reason: Success and then another dialog immediately says Error extracting ddc file. I'm not really sure what this means (certainly not success), but I've checked out the temporary file and it appears to be garbage data (although of the right file size). I don't know if anyone has an idea what might be done here to extract the files, or has dealt with a similar situation before. Any ideas are appreciated! Thank you in advance for your time and any help you can provide.

    Read the article

  • How can I make .vimrc read from an external file?

    - by GorillaSandwich
    I'd like to modify my .vimrc to read the value of a variable from an external file. How can I do this? Specifically, a friend and I share a git repo with our .vim files, but there are a few small differences in what we want in our configs. So most of the file is common, but we use if statements to determine whether to load user-specific sections, like this: let whoami = "user2" if whoami == "user1" ... After checking our common .vimrc out of source control, we each have to change the let whoami assignment so our own section will be loaded. Instead, I'd like to keep a separate file, which can be different for each of us, and from which vim will load that variable value. Maybe another angle on this is: Will vim automatically read all the files in my .vim directory? If so, we could each put a symlink in there called username.vim, and link that to an external file that would be different for each of us.

    Read the article

  • Can you resize all mac windows at once? (e.g. when docking)

    - by Ian Varley
    I've recently bought a Henge Dock, and like the ability to plug my MacBook in and instantly reconnect to all my peripherals (external monitor, keyboard, mouse, etc.). The only problem is that when I dock or undock, all of my windows are left at the wrong size (either too big or too small) and I have to manually resize them, one by one. Not the end of the world, but ... isn't there an easier way? I've seen Cinch and Sizeup, but they seem to only work on one window (the current app). I looked at the Automator as a possible way to do this, but it didn't seem to have any window operations. I also tried the AppleScript listed here, but it put the windows in wacky places.

    Read the article

  • DFS Replication, Users HOME folder - seems not to catch all files... any hints?

    - by TomTom
    I amm moving stuff out of a file Server. I am using DFS for that - the Folders are anyway in a DFS tree, so I can set up a replication temporarily, then drop the old Folder. Works nice, EXCEPT for the Folder containing the users home drives. Which, incidentally, is also the one I can not see all files in due to my permissions. Small Setup. We have 159mb in the users directories, 1280 files, 133 Folders original. The copy only has 157mb, 1269 files, 133 Folders. Anyone knwos of a way to find out what files are missing? IS this a Problem (could be some Caching files that are regenerated). Users are all offline (weekend) ;) This is pretty much the last share - all others had exactly ZERO issues.

    Read the article

  • What's faster, cp -R or unpacking tar.gz files?

    - by Buttle Butkus
    I have some tar.gz files that total many gigabytes on a CentOS system. Most of the tar.gz files are actually pretty small, but the ones with images are large. One is 7.7G, another is about 4G, and a couple around 1G. I have unpacked the files once already and now I want a second copy of all those files. I assumed that copying the unpacked files would be faster than re-unpacking them. But I started running cp -R about 10 minutes ago and so far less than 500M is copied. I feel certain that the unpacking process was faster. Am I right? And if so, why? It doesn't seem to make sense that unpacking would be faster than simply duplicating existing structures.

    Read the article

  • Is ceph usable with only 100Mbps bandwidth between nodes?

    - by vaab
    I haven't great hardware, but my requirements are low, I would like to start using ceph so as to abstract filesystem location and allow potential easy scaling to bigger hardware in an hypothetical future. My actual hardware meets ceph hardware requirements except the ethernet bandwidth part between the hosts. Mine is 100 Mbit/s which is much lower than the 1Gbps expected in ceph, even from the minimal requirement. Will I be able to use ceph in a very small smili-prod environnement (with limited number of clients) ? FYI: My hardware is 2 or maybe 3 hosts having each 4 core Intel, 24Go RAM, 2x2To disks but 100Mpbs between them.

    Read the article

  • New Servers Active Directory and Exchange

    - by user3164638
    I have 3 Dell PowerEdge server, each with 2 quad-core processors. I am going to bring this office out of the stone-age network (P2P, share files on a flash drive, emails through Google, etc) and set up Active Directory and Exchange 2013. Our needs are not that great at the moment - our staff consists of approximately 40 people, and our network may eventually be managed by an external company. We need only one domain for our emails (though we may serve emails for a few other partners domains as well). I was thinking of setting something up like this: Server 1: Primary DC. Active Directory and Exchange on separate virtual machines. Server 2: Redundant of server 1. Server 3: Shared resources, storage, backups, etc. How would you utilize 3 servers for an Active Directory / Exchange setup for a small/medium office? We do have plans to grow, so my solution must be scalable, though I'm not sure that I want to split permissions, though I'd consider it if that was something that could be changed on down the road.

    Read the article

  • Data unaccessible from WHS drive

    - by Eakraly
    Hi all, I had a Windows Home server machine that crashed. Before doing anything I decided to get data out of it. So I took the hard drive and connected it to Windows7 pc. What I get is that I cannot access almost any file! I do see directory structure, I can open small files like .txt and .ini but bigger files like .iso and video - no go. Same goes for Ubuntu and OS X - I can see files and even copy them - but they are corrupted. Any ideas for what the problem is?

    Read the article

  • Data unaccessible from WHS drive

    - by Eakraly
    Hi all, I had a Windows Home server machine that crashed. Before doing anything I decided to get data out of it. So I took the hard drive and connected it to Windows7 pc. What I get is that I cannot access almost any file! I do see directory structure, I can open small files like .txt and .ini but bigger files like .iso and video - no go. Same goes for Ubuntu and OS X - I can see files and even copy them - but they are corrupted. Any ideas for what the problem is?

    Read the article

  • Shell script to fix bad filenames? [closed]

    - by Ze'ev
    I'm IT at my small firm; and, despite my dire warnings, everyone puts files on the server with awful names, including leading & trailing spaces, bad characters (including \ ; / + . < > - etc!) They do this by accessing the (FreeBSD/FreeNAS) server via AFP on Macs, so no part of the system complains. Is there a script I can use to go through an entire directory tree and fix bad filenames? Basically replace all spaces & bad ASCII with _ ... and if a file already exists, just slap a _2 or something on the end. I don't suppose there's a way to get the system to enforce good filenaming conventions, is there?

    Read the article

  • How do I know if I need a level 3 switch?

    - by eekmeter
    We currently have a flat network with a bunch of unmanaged switches. I would like to use VLANs to segregate certain users like guests and I would like to use 802.1x. However, I'm not sure if what I need is a level 3 or a level 2 switch. From what I understand a level 3 switch does routing between VLANs. I don't think I need this at the moment but as I said I'm not sure since this is all new to me. What else would a level 3 switch do for me? Our network is relatively small, less than a 100 users. What exactly does a level 3 switch do that I can't get with a level 2 switch? When would I need a level 3?

    Read the article

< Previous Page | 250 251 252 253 254 255 256 257 258 259 260 261  | Next Page >