question
stringlengths 12
1.77k
| context
stringlengths 79
71.7k
| answer
stringlengths 1
1.63k
|
---|---|---|
Who is Kanad?
A. Kanad is Tendal 13 and Arvid 6's supervisor at the Ultroom.
B. Kanad is Reggie Laughton.
C. Kanad is the head of the whole galactic system.
D. Kanad is the leader of the Mycenae.
| Transcriber's Note: This etext was produced from Space Science Fiction May 1952. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed. THE ULTROOM ERROR by JERRY SOHL Smith admitted he had made an error involving a few murders—and a few thousand years. He was entitled to a sense of humor, though, even in the Ultroom! HB73782. Ultroom error. Tendal 13. Arvid 6. Kanad transfer out of 1609 complete, intact, but too near limit of 1,000 days. Next Kanad transfer ready. 1951. Reginald, son of Mr. and Mrs. Martin Laughton, 3495 Orland Drive, Marionville, Illinois, U. S. A. Arrive his 378th day. TB73782. Nancy Laughton sat on the blanket she had spread on the lawn in her front yard, knitting a pair of booties for the PTA bazaar. Occasionally she glanced at her son in the play pen, who was getting his daily dose of sunshine. He was gurgling happily, examining a ball, a cheese grater and a linen baby book, all with perfunctory interest. When she looked up again she noticed a man walking by—except he turned up the walk and crossed the lawn to her. He was a little taller than her husband, had piercing blue eyes and a rather amused set to his lips. "Hello, Nancy," he said. "Hello, Joe," she answered. It was her brother who lived in Kankakee. "I'm going to take the baby for a while," he said. "All right, Joe." He reached into the pen, picked up the baby. As he did so the baby's knees hit the side of the play pen and young Laughton let out a scream—half from hurt and half from sudden lack of confidence in his new handler. But this did not deter Joe. He started off with the child. Around the corner and after the man came a snarling mongrel dog, eyes bright, teeth glinting in the sunlight. The man did not turn as the dog threw himself at him, burying his teeth in his leg. Surprised, the man dropped the screaming child on the lawn and turned to the dog. Joe seemed off balance and he backed up confusedly in the face of the snapping jaws. Then he suddenly turned and walked away, the dog at his heels. "I tell you, the man said he was my brother and he made me think he was," Nancy told her husband for the tenth time. "I don't even have a brother." Martin Laughton sighed. "I can't understand why you believed him. It's just—just plain nuts, Nancy!" "Don't you think I know it?" Nancy said tearfully. "I feel like I'm going crazy. I can't say I dreamt it because there was Reggie with his bleeding knees, squalling for all he was worth on the grass—Oh, I don't even want to think about it." "We haven't lost Reggie, Nancy, remember that. Now why don't you try to get some rest?" "You—you don't believe me at all, do you, Martin?" When her husband did not answer, her head sank to her arms on the table and she sobbed. "Nancy, for heaven's sake, of course I believe you. I'm trying to think it out, that's all. We should have called the police." Nancy shook her head in her arms. "They'd—never—believe me either," she moaned. "I'd better go and make sure Reggie's all right." Martin got up out of his chair and went to the stairs. "I'm going with you," Nancy said, hurriedly rising and coming over to him. "We'll go up and look at him together." They found Reggie peacefully asleep in his crib in his room upstairs. They checked the windows and tucked in the blankets. They paused in the room for a moment and then Martin stole his arm around his wife and led her to the door. "As I've said, sergeant, this fellow hypnotized my wife. He made her think he was her brother. She doesn't even have a brother. Then he tried to get away with the baby." Martin leaned down and patted the dog. "It was Tiger here who scared him off." The police sergeant looked at the father, at Nancy and then at the dog. He scribbled notes in his book. "Are you a rich man, Mr. Laughton?" he asked. "Not at all. The bank still owns most of the house. I have a few hundred dollars, that's all." "What do you do?" "Office work, mostly. I'm a junior executive in an insurance company." "Any enemies?" "No ... Oh, I suppose I have a few people I don't get along with, like anybody else. Nobody who'd do anything like this, though." The sergeant flipped his notebook closed. "You'd better keep your dog inside and around the kid as much as possible. Keep your doors and windows locked. I'll see that the prowl car keeps an eye on the house. Call us if anything seems unusual or out of the way." Nancy had taken a sedative and was asleep by the time Martin finished cleaning the .30-.30 rifle he used for deer hunting. He put it by the stairs, ready for use, fully loaded, leaning it against the wall next to the telephone stand. The front door bell rang. He answered it. It was Dr. Stuart and another man. "I came as soon as I could, Martin," the young doctor said, stepping inside with the other man. "This is my new assistant, Dr. Tompkins." Martin and Tompkins shook hands. "The baby—?" Dr. Stuart asked. "Upstairs," Martin said. "You'd better get him, Dr. Tompkins, if we're to take him to the hospital. I'll stay here with Mr. Laughton. How've you been, Martin?" "Fine." "How's everything at the office?" "Fine." "And your wife?" "She's fine, too." "Glad to hear it, Martin. Mighty glad. Say, by the way, there's that bill you owe me. I think it's $32, isn't that right?" "Yes, I'd almost forgotten about it." "Why don't you be a good fellow and write a check for it? It's been over a year, you know." "That's right. I'll get right at it." Martin went over to his desk, opened it and started looking for his checkbook. Dr. Stuart stood by him, making idle comment until Dr. Tompkins came down the stairs with the sleeping baby cuddled against his shoulder. "Never mind the check, now, Martin. I see we're ready to go." He went over to his assistant and took the baby. Together they walked out the front door. "Good-bye," Martin said, going to the door. Then he was nearly bowled over by the discharge of the .30-.30. Dr. Stuart crumpled to the ground, the baby falling to the lawn. Dr. Tompkins whirled and there was a second shot. Dr. Tompkins pitched forward on his face. The figure of a woman ran from the house, retrieved the now squalling infant and ran back into the house. Once inside, Nancy slammed the door, gave the baby to the stunned Martin and headed for the telephone. "One of them was the same man!" she cried. Martin gasped, sinking into a chair with the baby. "I believed them," he said slowly and uncomprehendingly. "They made me believe them!" "Those bodies," the sergeant said. "Would you mind pointing them out to me, please?" "Aren't they—aren't they on the walk?" Mrs. Laughton asked. "There is nothing on the walk, Mrs. Laughton." "But there must be! I tell you I shot these men who posed as doctors. One of them was the same man who tried to take the baby this afternoon. They hypnotized my husband—" "Yes, I know, Mrs. Laughton. We've been through that." The sergeant went to the door and opened it. "Say, Homer, take another look around the walk and the bushes. There's supposed to be two of them. Shot with a .30-.30." He turned and picked up the gun and examined it again. "Ever shoot a gun before, Mrs. Laughton?" "Many times. Martin and I used to go hunting together before we had Reggie." The sergeant nodded. "You were taking an awful chance, shooting at a guy carrying your baby, don't you think?" "I shot him in the legs. The other—the other turned and I shot him in the chest. I could even see his eyes when he turned around. If I hadn't pulled the trigger then ... I don't want to remember it." The patrolman pushed the door open. "There's no bodies out here but there's some blood. Quite a lot of blood. A little to one side of the walk." The policemen went out. "Thank God you woke up, Nancy," Martin said. "I'd have let them have the baby." He reached over and smoothed the sleeping Reggie's hair. Nancy, who was rocking the boy, narrowed her eyes. "I wonder why they want our baby? He's just like any other baby. We don't have any money. We couldn't pay a ransom." "Reggie's pretty cute, though," Martin said. "You will have to admit that." Nancy smiled. Then she suddenly stopped rocking. "Martin!" He sat up quickly. "Where's Tiger?" Together they rose and walked around the room. They found him in a corner, eyes open, tongue protruding. He was dead. If we keep Reggie in the house much longer he'll turn out to be a hermit," Martin said at breakfast a month later. "He needs fresh air and sunshine." "I'm not going to sit on the lawn alone with him, Martin. I just can't, that's all. I'd be able to think of nothing but that day." "Still thinking about it? I think we'd have heard from them again if they were coming back. They probably got somebody else's baby by this time." Martin finished his coffee and rose to kiss her good-bye. "But for safety's sake I guess you'd better keep that gun handy." The morning turned into a brilliant, sunshiny day. Puffs of clouds moved slowly across the summer sky and a warm breeze rustled the trees. It would be a crime to keep Reggie inside on a day like this, Nancy thought. So she called Mrs. MacDougal, the next door neighbor. Mrs. MacDougal was familiar with what had happened to the Laughtons and she agreed to keep an eye on Nancy and Reggie and to call the police at the first sign of trouble. With a fearful but determined heart Nancy moved the play pen and set it up in the front yard. She spread a blanket for herself and put Reggie in the pen. Her heart pounded all the while and she watched the street for any strangers, ready to flee inside if need be. Reggie just gurgled with delight at the change in environment. This peaceful scene was disturbed by a speeding car in which two men were riding. The car roared up the street, swerved toward the parkway, tires screaming, bounced over the curb and sidewalk, straight toward the child and mother. Reggie, attracted by the sudden noise, looked up to see the approaching vehicle. His mother stood up, set her palms against her cheeks and shrieked. The car came on, crunched over the play pen, killing the child. The mother was hit and instantly killed, force of the blow snapping her spine and tossing her against the house. The car plunged on into a tree, hitting it a terrible blow, crumbling the car's forward end so it looked like an accordion. The men were thrown from the machine. "We'll never be able to prosecute in this case," the states attorney said. "At least not on a drunken driving basis." "I can't get over it," the chief of police said. "I've got at least six men who will swear the man was drunk. He staggered, reeled and gave the usual drunk talk. He reeked of whiskey." The prosecutor handed the report over the desk. "Here's the analysis. Not a trace of alcohol. He couldn't have even had a smell of near beer. Here's another report. This is his physical exam made not long afterwards. The man was in perfect health. Only variations are he had a scar on his leg where something, probably a dog, bit him once. And then a scar on his chest. It looked like an old gunshot wound, they said. Must have happened years ago." "That's odd. The man who accosted Mrs. Laughton in the afternoon was bitten by their dog. Later that night she said she shot the same man in the chest. Since the scars are healed it obviously couldn't be the same man. But there's a real coincidence for you. And speaking of the dogbite, the Laughton dog died that night. His menu evidently didn't agree with him. Never did figure what killed him, actually." "Any record of treatment on the man she shot?" "The men . You'll remember, there were two. No, we never found a trace of either. No doctor ever made a report of a gunshot wound that night. No hospital had a case either—at least not within several hundred miles—that night or several nights afterwards. Ever been shot with .30-.30?" The state attorney shook his head. "I wouldn't be here if I had." "I'll say you wouldn't. The pair must have crawled away to die God knows where." "Getting back to the man who ran over the child and killed Mrs. Laughton. Why did he pretend to be drunk?" It was the chief's turn to shake his head. "Your guess is as good as mine. There are a lot of angles to this case none of us understand. It looks deliberate, but where's the motive?" "What does the man have to say?" "I was afraid you'd get to him," the chief said, his neck reddening. "It's all been rather embarrassing to the department." He coughed self-consciously. "He's proved a strange one, all right. He says his name is John Smith and he's got cards to prove it, too—for example, a social security card. It looks authentic, yet there's no such number on file in Washington, so we've discovered. We've had him in jail for a week and we've all taken turns questioning him. He laughs and admits his guilt—in fact, he seems amused by most everything. Sometimes all alone in his cell he'll start laughing for no apparent reason. It gives you the creeps." The states attorney leaned back in his chair. "Maybe it's a case for an alienist." "One jump ahead of you. Dr. Stone thinks he's normal, but won't put down any I.Q. Actually, he can't figure him out himself. Smith seems to take delight in answering questions—sort of anticipates them and has the answer ready before you're half through asking." "Well, if Dr. Stone says he's normal, that's enough for me." The prosecutor was silent for a moment. Then, "How about the husband?" "Laughton? We're afraid to let him see him. All broken up. No telling what kind of a rumpus he'd start—especially if Smith started his funny business." "Guess you're right. Well, Mr. Smith won't think it's so funny when we hang criminal negligence or manslaughter on him. By the way, you've checked possible family connections?" "Nobody ever saw John Smith before. Even at the address on his driver's license. And there's no duplicate of that in Springfield, in case you're interested." The man who had laughingly told police his name was John Smith lay on his cot in the county jail, his eyes closed, his arms folded across his chest. This gave him the appearance of being alert despite reclining. Even as he lay, his mouth held a hint of a smile. Arvid 6—for John Smith was Arvid 6—had lain in that position for more than four hours, when suddenly he snapped his eyes open and appeared to be listening. For a moment a look of concern crossed his face and he swung his legs to the floor and sat there expectantly. Arvid 6 knew Tendal 13 had materialized and was somewhere in the building. Eventually there were some sounds from beyond the steel cell and doorway. There was a clang when the outer doorway was opened and Arvid 6 rose from his cot. "Your lawyer's here to see you," the jailer said, indicating the man with the brief case. "Ring the buzzer when you're through." The jailer let the man in, locked the cell door and walked away. The man threw the brief case on the jail cot and stood glaring. "Your damned foolishness has gone far enough. I'm sick and tired of it," he declared. "If you carry on any more we'll never get back to the Ultroom!" "I'm sorry, Tendal," the man on the cot said. "I didn't think—" "You're absolutely right. You didn't think. Crashing that car into that tree and killing that woman—that was the last straw. You don't even deserve to get back to our era. You ought to be made to rot here." "I'm really sorry about that," Arvid 6 said. You know the instructions. Just because you work in the Ultroom don't get to thinking human life doesn't have any value. We wouldn't be here if it hadn't. But to unnecessarily kill—" The older man shook his head. "You could have killed yourself as well and we'd never get the job done. As it is, you almost totally obliterated me." Tendal 13 paced the length of the cell and back again, gesturing as he talked. "It was only with the greatest effort I pulled myself back together again. I doubt that you could have done it. And then all the while you've been sitting here, probably enjoying yourself with your special brand of humor I have grown to despise." "You didn't have to come along at all, you know," Arvid 6 said. "How well I know! How sorry I am that I ever did! It was only because I was sorry for you, because someone older and more experienced than you was needed. I volunteered. Imagine that! I volunteered! Tendal 13 reaches the height of stupidity and volunteers to help Arvid 6 go back 6,000 years to bring Kanad back, to correct a mistake Arvid 6 made!" He snorted. "I still can't believe I was ever that stupid. I only prove it when I pinch myself and here I am. "Oh, you've been a joy to be with! First it was that hunt in ancient Mycenae when you let the lion escape the hunters' quaint spears and we were partly eaten by the lion in the bargain, although you dazzled the hunters, deflecting their spears. And then your zest for drink when we were with Octavian in Alexandria that led to everybody's amusement but ours when we were ambushed by Anthony's men. And worst of all, that English barmaid you became engrossed with at our last stop in 1609, when her husband mistook me for you and you let him take me apart piece by piece—" "All right, all right," Arvid 6 said. "I'll admit I've made some mistakes. You're just not adventurous, that's all." "Shut up! For once you're going to listen to me. Our instructions specifically stated we were to have as little as possible to do with these people. But at every turn you've got us more and more enmeshed with them. If that's adventure, you can have it." Tendal 13 sat down wearily and sank his head in his hands. "It was you who conceived the idea of taking Reggie right out of his play pen. 'Watch me take that child right out from under its mother's nose' were your exact words. And before I could stop you, you did. Only you forgot an important factor in the equation—the dog, Tiger. And you nursed a dogbite most of the afternoon before it healed. And then you took your spite out on the poor thing by suggesting suffocation to it that night. "And speaking of that night, you remember we agreed I was to do the talking. But no, you pulled a switch and captured Martin Laughton's attention. 'I came as soon as I could, Martin,' you said. And suddenly I played a very minor role. 'This is my new assistant, Dr. Tompkins,' you said. And then what happened? I get shot in the legs and you get a hole in your back. We were both nearly obliterated that time and we didn't even come close to getting the child. "Still you wanted to run the whole show. 'I'm younger than you,' you said. 'I'll take the wheel.' And the next thing I know I'm floating in space halfway to nowhere with two broken legs, a spinal injury, concussion and some of the finest bruises you ever saw." These twentieth century machines aren't what they ought to be," Arvid 6 said. "You never run out of excuses, do you, Arvid? Remember what you said in the Ultroom when you pushed the lever clear over and transferred Kanad back 6,000 years? 'My hand slipped.' As simple as that. 'My hand slipped.' It was so simple everyone believed you. You were given no real punishment. In a way it was a reward—at least to you—getting to go back and rescue the life germ of Kanad out of each era he'd be born in." Tendal 13 turned and looked steadily and directly at Arvid 6. "Do you know what I think? I think you deliberately pushed the lever over as far as it would go just to see what would happen . That's how simple I think it was." Arvid 6 flushed, turned away and looked at the floor. "What crazy things have you been doing since I've been gone?" Tendal 13 asked. Arvid 6 sighed. "After what you just said I guess it wouldn't amuse you, although it has me. They got to me right after the accident before I had a chance to collect my wits, dematerialize or anything—you said we shouldn't dematerialize in front of anybody." "That's right." "Well, I didn't know what to do. I could see they thought I was drunk, so I was. But they had a blood sample before I could manufacture any alcohol in my blood, although I implanted a memory in them that I reeked of it." He laughed. "I fancy they're thoroughly confused." "And you're thoroughly amused, no doubt. Have they questioned you?" "At great length. They had a psychiatrist in to see me. He was a queer fellow with the most stupid set of questions and tests I ever saw." "And you amused yourself with him." "I suppose you'd think so." "Who do you tell them you are?" "John Smith. A rather prevalent name here, I understand. I manufactured a pasteboard called a social security card and a driver's license—" "Never mind. It's easy to see you've been your own inimitable self. Believe me, if I ever get back to the Ultroom I hope I never see you again. And I hope I'll never leave there again though I'm rejuvenated through a million years." "Was Kanad's life germ transferred all right this time?" Tendal 13 shook his head. "I haven't heard. The transfers are getting more difficult all the time. In 1609, you'll remember, it was a case of pneumonia for the two-year-old. A simple procedure. It wouldn't work here. Medicine's too far along." He produced a notebook. "The last jump was 342 years, a little more than average. The next ought to be around 2250. Things will be more difficult than ever there, probably." "Do you think Kanad will be angry about all this?" "How would you like to have to go through all those birth processes, to have your life germ knocked from one era to the next?" "Frankly, I didn't think he'd go back so far." "If it had been anybody but Kanad nobody'd ever have thought of going back after it. The life germ of the head of the whole galactic system who came to the Ultroom to be transplanted to a younger body—and then sending him back beyond his original birth date—" Tendal 13 got up and commenced his pacing again. "Oh, I suppose Kanad's partly to blame, wanting rejuvenating at only 300 years. Some have waited a thousand or more or until their bones are like paper." "I just wonder how angry Kanad will be," Arvid muttered. HB92167. Ultroom Error. Tendal 13. Arvid 6. Kanad transfer out of 1951 complete. Next Kanad transfer ready. 2267. Phullam 19, son of Orla 39 and Rhoda R, 22H Level M, Hemisphere B, Quadrant 3, Sector I. Arrive his 329th Day. TB92167 Arvid 6 rose from the cot and the two men faced each other. "Before we leave, Arvid," Tendal 13 started to say. "I know, I know. You want me to let you handle everything." "Exactly. Is that too much to ask after all you've done?" "I guess I have made mistakes. From now on you be the boss. I'll do whatever you say." "I hope I can count on that." Tendal 13 rang the jail buzzer. The jailer unlocked the cell door. "You remember the chief said it's all right to take him with me, Matthews," Tendal 13 told the jailer. "Yes, I remember," the jailer said mechanically, letting them both out of the cell. They walked together down the jail corridor. When they came to another barred door the jailer fumbled with the keys and clumsily tried several with no luck. Arvid 6, an amused set to his mouth and devilment in his eyes, watched the jailer's expression as he walked through the bars of the door. He laughed as he saw the jailer's eyes bulge. "Arvid!" Tendal 13 walked briskly through the door, snatched Arvid 6 by the shoulders and shook him. The jailer watched stupified as the two men vanished in the middle of a violent argument. | C. Kanad is the head of the whole galactic system. |
Why is the Geig Corps important?
A. Val and Ron worked for them before signing up with UranCo
B. It is UranCo's method of acquiring manpower for the resource search
C. It is how Ledman got involved in the uranium project in the first place
D. They funded the dome that Ledman lives in
| THE HUNTED HEROES By ROBERT SILVERBERG The planet itself was tough enough—barren, desolate, forbidding; enough to stop the most adventurous and dedicated. But they had to run head-on against a mad genius who had a motto: Death to all Terrans! "Let's keep moving," I told Val. "The surest way to die out here on Mars is to give up." I reached over and turned up the pressure on her oxymask to make things a little easier for her. Through the glassite of the mask, I could see her face contorted in an agony of fatigue. And she probably thought the failure of the sandcat was all my fault, too. Val's usually about the best wife a guy could ask for, but when she wants to be she can be a real flying bother. It was beyond her to see that some grease monkey back at the Dome was at fault—whoever it was who had failed to fasten down the engine hood. Nothing but what had stopped us could stop a sandcat: sand in the delicate mechanism of the atomic engine. But no; she blamed it all on me somehow: So we were out walking on the spongy sand of the Martian desert. We'd been walking a good eight hours. "Can't we turn back now, Ron?" Val pleaded. "Maybe there isn't any uranium in this sector at all. I think we're crazy to keep on searching out here!" I started to tell her that the UranCo chief had assured me we'd hit something out this way, but changed my mind. When Val's tired and overwrought there's no sense in arguing with her. I stared ahead at the bleak, desolate wastes of the Martian landscape. Behind us somewhere was the comfort of the Dome, ahead nothing but the mazes and gullies of this dead world. He was a cripple in a wheelchair—helpless as a rattlesnake. "Try to keep going, Val." My gloved hand reached out and clumsily enfolded hers. "Come on, kid. Remember—we're doing this for Earth. We're heroes." She glared at me. "Heroes, hell!" she muttered. "That's the way it looked back home, but, out there it doesn't seem so glorious. And UranCo's pay is stinking." "We didn't come out here for the pay, Val." "I know, I know, but just the same—" It must have been hell for her. We had wandered fruitlessly over the red sands all day, both of us listening for the clicks of the counter. And the geigers had been obstinately hushed all day, except for their constant undercurrent of meaningless noises. Even though the Martian gravity was only a fraction of Earth's, I was starting to tire, and I knew it must have been really rough on Val with her lovely but unrugged legs. "Heroes," she said bitterly. "We're not heroes—we're suckers! Why did I ever let you volunteer for the Geig Corps and drag me along?" Which wasn't anywhere close to the truth. Now I knew she was at the breaking point, because Val didn't lie unless she was so exhausted she didn't know what she was doing. She had been just as much inflamed by the idea of coming to Mars to help in the search for uranium as I was. We knew the pay was poor, but we had felt it a sort of obligation, something we could do as individuals to keep the industries of radioactives-starved Earth going. And we'd always had a roving foot, both of us. No, we had decided together to come to Mars—the way we decided together on everything. Now she was turning against me. I tried to jolly her. "Buck up, kid," I said. I didn't dare turn up her oxy pressure any higher, but it was obvious she couldn't keep going. She was almost sleep-walking now. We pressed on over the barren terrain. The geiger kept up a fairly steady click-pattern, but never broke into that sudden explosive tumult that meant we had found pay-dirt. I started to feel tired myself, terribly tired. I longed to lie down on the soft, spongy Martian sand and bury myself. I looked at Val. She was dragging along with her eyes half-shut. I felt almost guilty for having dragged her out to Mars, until I recalled that I hadn't. In fact, she had come up with the idea before I did. I wished there was some way of turning the weary, bedraggled girl at my side back into the Val who had so enthusiastically suggested we join the Geigs. Twelve steps later, I decided this was about as far as we could go. I stopped, slipped out of the geiger harness, and lowered myself ponderously to the ground. "What'samatter, Ron?" Val asked sleepily. "Something wrong?" "No, baby," I said, putting out a hand and taking hers. "I think we ought to rest a little before we go any further. It's been a long, hard day." It didn't take much to persuade her. She slid down beside me, curled up, and in a moment she was fast asleep, sprawled out on the sands. Poor kid , I thought. Maybe we shouldn't have come to Mars after all. But, I reminded myself, someone had to do the job. A second thought appeared, but I squelched it: Why the hell me? I looked down at Valerie's sleeping form, and thought of our warm, comfortable little home on Earth. It wasn't much, but people in love don't need very fancy surroundings. I watched her, sleeping peacefully, a wayward lock of her soft blonde hair trailing down over one eyebrow, and it seemed hard to believe that we'd exchanged Earth and all it held for us for the raw, untamed struggle that was Mars. But I knew I'd do it again, if I had the chance. It's because we wanted to keep what we had. Heroes? Hell, no. We just liked our comforts, and wanted to keep them. Which took a little work. Time to get moving. But then Val stirred and rolled over in her sleep, and I didn't have the heart to wake her. I sat there, holding her, staring out over the desert, watching the wind whip the sand up into weird shapes. The Geig Corps preferred married couples, working in teams. That's what had finally decided it for us—we were a good team. We had no ties on Earth that couldn't be broken without much difficulty. So we volunteered. And here we are. Heroes. The wind blasted a mass of sand into my face, and I felt it tinkle against the oxymask. I glanced at the suit-chronometer. Getting late. I decided once again to wake Val. But she was tired. And I was tired too, tired from our wearying journey across the empty desert. I started to shake Val. But I never finished. It would be so nice just to lean back and nuzzle up to her, down in the sand. So nice. I yawned, and stretched back. I awoke with a sudden startled shiver, and realized angrily I had let myself doze off. "Come on, Val," I said savagely, and started to rise to my feet. I couldn't. I looked down. I was neatly bound in thin, tough, plastic tangle-cord, swathed from chin to boot-bottoms, my arms imprisoned, my feet caught. And tangle-cord is about as easy to get out of as a spider's web is for a trapped fly. It wasn't Martians that had done it. There weren't any Martians, hadn't been for a million years. It was some Earthman who had bound us. I rolled my eyes toward Val, and saw that she was similarly trussed in the sticky stuff. The tangle-cord was still fresh, giving off a faint, repugnant odor like that of drying fish. It had been spun on us only a short time ago, I realized. "Ron—" "Don't try to move, baby. This stuff can break your neck if you twist it wrong." She continued for a moment to struggle futilely, and I had to snap, "Lie still, Val!" "A very wise statement," said a brittle, harsh voice from above me. I looked up and saw a helmeted figure above us. He wasn't wearing the customary skin-tight pliable oxysuits we had. He wore an outmoded, bulky spacesuit and a fishbowl helmet, all but the face area opaque. The oxygen cannisters weren't attached to his back as expected, though. They were strapped to the back of the wheelchair in which he sat. Through the fishbowl I could see hard little eyes, a yellowed, parchment-like face, a grim-set jaw. I didn't recognize him, and this struck me odd. I thought I knew everyone on sparsely-settled Mars. Somehow I'd missed him. What shocked me most was that he had no legs. The spacesuit ended neatly at the thighs. He was holding in his left hand the tanglegun with which he had entrapped us, and a very efficient-looking blaster was in his right. "I didn't want to disturb your sleep," he said coldly. "So I've been waiting here for you to wake up." I could just see it. He might have been sitting there for hours, complacently waiting to see how we'd wake up. That was when I realized he must be totally insane. I could feel my stomach-muscles tighten, my throat constrict painfully. Then anger ripped through me, washing away the terror. "What's going on?" I demanded, staring at the half of a man who confronted us from the wheelchair. "Who are you?" "You'll find out soon enough," he said. "Suppose now you come with me." He reached for the tanglegun, flipped the little switch on its side to MELT, and shot a stream of watery fluid over our legs, keeping the blaster trained on us all the while. Our legs were free. "You may get up now," he said. "Slowly, without trying to make trouble." Val and I helped each other to our feet as best we could, considering our arms were still tightly bound against the sides of our oxysuits. "Walk," the stranger said, waving the tanglegun to indicate the direction. "I'll be right behind you." He holstered the tanglegun. I glimpsed the bulk of an outboard atomic rigging behind him, strapped to the back of the wheelchair. He fingered a knob on the arm of the chair and the two exhaust ducts behind the wheel-housings flamed for a moment, and the chair began to roll. Obediently, we started walking. You don't argue with a blaster, even if the man pointing it is in a wheelchair. "What's going on, Ron?" Val asked in a low voice as we walked. Behind us the wheelchair hissed steadily. "I don't quite know, Val. I've never seen this guy before, and I thought I knew everyone at the Dome." "Quiet up there!" our captor called, and we stopped talking. We trudged along together, with him following behind; I could hear the crunch-crunch of the wheelchair as its wheels chewed into the sand. I wondered where we were going, and why. I wondered why we had ever left Earth. The answer to that came to me quick enough: we had to. Earth needed radioactives, and the only way to get them was to get out and look. The great atomic wars of the late 20th Century had used up much of the supply, but the amount used to blow up half the great cities of the world hardly compared with the amount we needed to put them back together again. In three centuries the shattered world had been completely rebuilt. The wreckage of New York and Shanghai and London and all the other ruined cities had been hidden by a shining new world of gleaming towers and flying roadways. We had profited by our grandparents' mistakes. They had used their atomics to make bombs. We used ours for fuel. It was an atomic world. Everything: power drills, printing presses, typewriters, can openers, ocean liners, powered by the inexhaustible energy of the dividing atom. But though the energy is inexhaustible, the supply of nuclei isn't. After three centuries of heavy consumption, the supply failed. The mighty machine that was Earth's industry had started to slow down. And that started the chain of events that led Val and me to end up as a madman's prisoners, on Mars. With every source of uranium mined dry on Earth, we had tried other possibilities. All sorts of schemes came forth. Project Sea-Dredge was trying to get uranium from the oceans. In forty or fifty years, they'd get some results, we hoped. But there wasn't forty or fifty years' worth of raw stuff to tide us over until then. In a decade or so, our power would be just about gone. I could picture the sort of dog-eat-dog world we'd revert back to. Millions of starving, freezing humans tooth-and-clawing in it in the useless shell of a great atomic civilization. So, Mars. There's not much uranium on Mars, and it's not easy to find or any cinch to mine. But what little is there, helps. It's a stopgap effort, just to keep things moving until Project Sea-Dredge starts functioning. Enter the Geig Corps: volunteers out on the face of Mars, combing for its uranium deposits. And here we are, I thought. After we walked on a while, a Dome became visible up ahead. It slid up over the crest of a hill, set back between two hummocks on the desert. Just out of the way enough to escape observation. For a puzzled moment I thought it was our Dome, the settlement where all of UranCo's Geig Corps were located, but another look told me that this was actually quite near us and fairly small. A one-man Dome, of all things! "Welcome to my home," he said. "The name is Gregory Ledman." He herded us off to one side of the airlock, uttered a few words keyed to his voice, and motioned us inside when the door slid up. When we were inside he reached up, clumsily holding the blaster, and unscrewed the ancient spacesuit fishbowl. His face was a bitter, dried-up mask. He was a man who hated. The place was spartanly furnished. No chairs, no tape-player, no decoration of any sort. Hard bulkhead walls, rivet-studded, glared back at us. He had an automatic chef, a bed, and a writing-desk, and no other furniture. Suddenly he drew the tanglegun and sprayed our legs again. We toppled heavily to the floor. I looked up angrily. "I imagine you want to know the whole story," he said. "The others did, too." Valerie looked at me anxiously. Her pretty face was a dead white behind her oxymask. "What others?" "I never bothered to find out their names," Ledman said casually. "They were other Geigs I caught unawares, like you, out on the desert. That's the only sport I have left—Geig-hunting. Look out there." He gestured through the translucent skin of the Dome, and I felt sick. There was a little heap of bones lying there, looking oddly bright against the redness of the sands. They were the dried, parched skeletons of Earthmen. Bits of cloth and plastic, once oxymasks and suits, still clung to them. Suddenly I remembered. There had been a pattern there all the time. We didn't much talk about it; we chalked it off as occupational hazards. There had been a pattern of disappearances on the desert. I could think of six, eight names now. None of them had been particularly close friends. You don't get time to make close friends out here. But we'd vowed it wouldn't happen to us. It had. "You've been hunting Geigs?" I asked. " Why? What've they ever done to you?" He smiled, as calmly as if I'd just praised his house-keeping. "Because I hate you," he said blandly. "I intend to wipe every last one of you out, one by one." I stared at him. I'd never seen a man like this before; I thought all his kind had died at the time of the atomic wars. I heard Val sob, "He's a madman!" "No," Ledman said evenly. "I'm quite sane, believe me. But I'm determined to drive the Geigs—and UranCo—off Mars. Eventually I'll scare you all away." "Just pick us off in the desert?" "Exactly," replied Ledman. "And I have no fears of an armed attack. This place is well fortified. I've devoted years to building it. And I'm back against those hills. They couldn't pry me out." He let his pale hand run up into his gnarled hair. "I've devoted years to this. Ever since—ever since I landed here on Mars." "What are you going to do with us?" Val finally asked, after a long silence. He didn't smile this time. "Kill you," he told her. "Not your husband. I want him as an envoy, to go back and tell the others to clear off." He rocked back and forth in his wheelchair, toying with the gleaming, deadly blaster in his hand. We stared in horror. It was a nightmare—sitting there, placidly rocking back and forth, a nightmare. I found myself fervently wishing I was back out there on the infinitely safer desert. "Do I shock you?" he asked. "I shouldn't—not when you see my motives." "We don't see them," I snapped. "Well, let me show you. You're on Mars hunting uranium, right? To mine and ship the radioactives back to Earth to keep the atomic engines going. Right?" I nodded over at our geiger counters. "We volunteered to come to Mars," Val said irrelevantly. "Ah—two young heroes," Ledman said acidly. "How sad. I could almost feel sorry for you. Almost." "Just what is it you're after?" I said, stalling, stalling. "Atomics cost me my legs," he said. "You remember the Sadlerville Blast?" he asked. "Of course." And I did, too. I'd never forget it. No one would. How could I forget that great accident—killing hundreds, injuring thousands more, sterilizing forty miles of Mississippi land—when the Sadlerville pile went up? "I was there on business at the time," Ledman said. "I represented Ledman Atomics. I was there to sign a new contract for my company. You know who I am, now?" I nodded. "I was fairly well shielded when it happened. I never got the contract, but I got a good dose of radiation instead. Not enough to kill me," he said. "Just enough to necessitate the removal of—" he indicated the empty space at his thighs. "So I got off lightly." He gestured at the wheelchair blanket. I still didn't understand. "But why kill us Geigs? We had nothing to do with it." "You're just in this by accident," he said. "You see, after the explosion and the amputation, my fellow-members on the board of Ledman Atomics decided that a semi-basket case like myself was a poor risk as Head of the Board, and they took my company away. All quite legal, I assure you. They left me almost a pauper!" Then he snapped the punchline at me. "They renamed Ledman Atomics. Who did you say you worked for?" I began, "Uran—" "Don't bother. A more inventive title than Ledman Atomics, but not quite as much heart, wouldn't you say?" He grinned. "I saved for years; then I came to Mars, lost myself, built this Dome, and swore to get even. There's not a great deal of uranium on this planet, but enough to keep me in a style to which, unfortunately, I'm no longer accustomed." He consulted his wrist watch. "Time for my injection." He pulled out the tanglegun and sprayed us again, just to make doubly certain. "That's another little souvenir of Sadlerville. I'm short on red blood corpuscles." He rolled over to a wall table and fumbled in a container among a pile of hypodermics. "There are other injections, too. Adrenalin, insulin. Others. The Blast turned me into a walking pin-cushion. But I'll pay it all back," he said. He plunged the needle into his arm. My eyes widened. It was too nightmarish to be real. I wasn't seriously worried about his threat to wipe out the entire Geig Corps, since it was unlikely that one man in a wheelchair could pick us all off. No, it wasn't the threat that disturbed me, so much as the whole concept, so strange to me, that the human mind could be as warped and twisted as Ledman's. I saw the horror on Val's face, and I knew she felt the same way I did. "Do you really think you can succeed?" I taunted him. "Really think you can kill every Earthman on Mars? Of all the insane, cockeyed—" Val's quick, worried head-shake cut me off. But Ledman had felt my words, all right. "Yes! I'll get even with every one of you for taking away my legs! If we hadn't meddled with the atom in the first place, I'd be as tall and powerful as you, today—instead of a useless cripple in a wheelchair." "You're sick, Gregory Ledman," Val said quietly. "You've conceived an impossible scheme of revenge and now you're taking it out on innocent people who've done nothing, nothing at all to you. That's not sane!" His eyes blazed. "Who are you to talk of sanity?" Uneasily I caught Val's glance from a corner of my eye. Sweat was rolling down her smooth forehead faster than the auto-wiper could swab it away. "Why don't you do something? What are you waiting for, Ron?" "Easy, baby," I said. I knew what our ace in the hole was. But I had to get Ledman within reach of me first. "Enough," he said. "I'm going to turn you loose outside, right after—" " Get sick! " I hissed to Val, low. She began immediately to cough violently, emitting harsh, choking sobs. "Can't breathe!" She began to yell, writhing in her bonds. That did it. Ledman hadn't much humanity left in him, but there was a little. He lowered the blaster a bit and wheeled one-hand over to see what was wrong with Val. She continued to retch and moan most horribly. It almost convinced me. I saw Val's pale, frightened face turn to me. He approached and peered down at her. He opened his mouth to say something, and at that moment I snapped my leg up hard, tearing the tangle-cord with a snicking rasp, and kicked his wheelchair over. The blaster went off, burning a hole through the Dome roof. The automatic sealers glued-in instantly. Ledman went sprawling helplessly out into the middle of the floor, the wheelchair upended next to him, its wheels slowly revolving in the air. The blaster flew from his hands at the impact of landing and spun out near me. In one quick motion I rolled over and covered it with my body. Ledman clawed his way to me with tremendous effort and tried wildly to pry the blaster out from under me, but without success. I twisted a bit, reached out with my free leg, and booted him across the floor. He fetched up against the wall of the Dome and lay there. Val rolled over to me. "Now if I could get free of this stuff," I said, "I could get him covered before he comes to. But how?" "Teamwork," Val said. She swivelled around on the floor until her head was near my boot. "Push my oxymask off with your foot, if you can." I searched for the clamp and tried to flip it. No luck, with my heavy, clumsy boot. I tried again, and this time it snapped open. I got the tip of my boot in and pried upward. The oxymask came off, slowly, scraping a jagged red scratch up the side of Val's neck as it came. "There," she breathed. "That's that." I looked uneasily at Ledman. He was groaning and beginning to stir. Val rolled on the floor and her face lay near my right arm. I saw what she had in mind. She began to nibble the vile-tasting tangle-cord, running her teeth up and down it until it started to give. She continued unfailingly. Finally one strand snapped. Then another. At last I had enough use of my hand to reach out and grasp the blaster. Then I pulled myself across the floor to Ledman, removed the tanglegun, and melted the remaining tangle-cord off. My muscles were stiff and bunched, and rising made me wince. I turned and freed Val. Then I turned and faced Ledman. "I suppose you'll kill me now," he said. "No. That's the difference between sane people and insane," I told him. "I'm not going to kill you at all. I'm going to see to it that you're sent back to Earth." " No! " he shouted. "No! Anything but back there. I don't want to face them again—not after what they did to me—" "Not so loud," I broke in. "They'll help you on Earth. They'll take all the hatred and sickness out of you, and turn you into a useful member of society again." "I hate Earthmen," he spat out. "I hate all of them." "I know," I said sarcastically. "You're just all full of hate. You hated us so much that you couldn't bear to hang around on Earth for as much as a year after the Sadlerville Blast. You had to take right off for Mars without a moment's delay, didn't you? You hated Earth so much you had to leave." "Why are you telling all this to me?" "Because if you'd stayed long enough, you'd have used some of your pension money to buy yourself a pair of prosthetic legs, and then you wouldn't need this wheelchair." Ledman scowled, and then his face went belligerent again. "They told me I was paralyzed below the waist. That I'd never walk again, even with prosthetic legs, because I had no muscles to fit them to." "You left Earth too quickly," Val said. "It was the only way," he protested. "I had to get off—" "She's right," I told him. "The atom can take away, but it can give as well. Soon after you left they developed atomic-powered prosthetics—amazing things, virtually robot legs. All the survivors of the Sadlerville Blast were given the necessary replacement limbs free of charge. All except you. You were so sick you had to get away from the world you despised and come here." "You're lying," he said. "It's not true!" "Oh, but it is," Val smiled. I saw him wilt visibly, and for a moment I almost felt sorry for him, a pathetic legless figure propped up against the wall of the Dome at blaster-point. But then I remembered he'd killed twelve Geigs—or more—and would have added Val to the number had he had the chance. "You're a very sick man, Ledman," I said. "All this time you could have been happy, useful on Earth, instead of being holed up here nursing your hatred. You might have been useful, on Earth. But you decided to channel everything out as revenge." "I still don't believe it—those legs. I might have walked again. No—no, it's all a lie. They told me I'd never walk," he said, weakly but stubbornly still. I could see his whole structure of hate starting to topple, and I decided to give it the final push. "Haven't you wondered how I managed to break the tangle-cord when I kicked you over?" "Yes—human legs aren't strong enough to break tangle-cord that way." "Of course not," I said. I gave Val the blaster and slipped out of my oxysuit. "Look," I said. I pointed to my smooth, gleaming metal legs. The almost soundless purr of their motors was the only noise in the room. "I was in the Sadlerville Blast, too," I said. "But I didn't go crazy with hate when I lost my legs." Ledman was sobbing. "Okay, Ledman," I said. Val got him into his suit, and brought him the fishbowl helmet. "Get your helmet on and let's go. Between the psychs and the prosthetics men, you'll be a new man inside of a year." "But I'm a murderer!" "That's right. And you'll be sentenced to psych adjustment. When they're finished, Gregory Ledman the killer will be as dead as if they'd electrocuted you, but there'll be a new—and sane—Gregory Ledman." I turned to Val. "Got the geigers, honey?" For the first time since Ledman had caught us, I remembered how tired Val had been out on the desert. I realized now that I had been driving her mercilessly—me, with my chromium legs and atomic-powered muscles. No wonder she was ready to fold! And I'd been too dense to see how unfair I had been. She lifted the geiger harnesses, and I put Ledman back in his wheelchair. Val slipped her oxymask back on and fastened it shut. "Let's get back to the Dome in a hurry," I said. "We'll turn Ledman over to the authorities. Then we can catch the next ship for Earth." "Go back? Go back? If you think I'm backing down now and quitting you can find yourself another wife! After we dump this guy I'm sacking in for twenty hours, and then we're going back out there to finish that search-pattern. Earth needs uranium, honey, and I know you'd never be happy quitting in the middle like that." She smiled. "I can't wait to get out there and start listening for those tell-tale clicks." I gave a joyful whoop and swung her around. When I put her down, she squeezed my hand, hard. "Let's get moving, fellow hero," she said. I pressed the stud for the airlock, smiling. THE END Transcriber's Note: This etext was produced from Amazing Stories September 1956. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed. Minor spelling and typographical errors have been corrected without note. | B. It is UranCo's method of acquiring manpower for the resource search |
How was Johnson convinced to buy the case astroid fever medication?
A. Proven statistics showing that it was the best antidote
B. Joe's acting skills
C. He felt feverish and thought he may have contracted the illness
D. A price too good that could not be turned down
| GRIFTERS' ASTEROID By H. L. GOLD Harvey and Joe were the slickest con-men ever to gyp a space-lane sucker. Or so they thought! Angus Johnson knew differently. He charged them five buckos for a glass of water—and got it! [Transcriber's Note: This etext was produced from Planet Stories May 1943. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed.] Characteristically, Harvey Ellsworth tried to maintain his dignity, though his parched tongue was almost hanging out. But Joe Mallon, with no dignity to maintain, lurched across the rubbish-strewn patch of land that had been termed a spaceport. When Harvey staggered pontifically into the battered metalloy saloon—the only one on Planetoid 42—his tall, gangling partner was already stumbling out, mouthing something incoherent. They met in the doorway, violently. "We're delirious!" Joe cried. "It's a mirage!" "What is?" asked Harvey through a mouthful of cotton. Joe reeled aside, and Harvey saw what had upset his partner. He stared, speechless for once. In their hectic voyages from planet to planet, the pair of panacea purveyors had encountered the usual strange life-forms. But never had they seen anything like the amazing creature in that colonial saloon. Paying no attention to them, it was carrying a case of liquor in two hands, six siphons in two others, and a broom and dustpan in the remaining pair. The bartender, a big man resembling the plumpish Harvey in build, was leaning negligently on the counter, ordering this impossible being to fill the partly-emptied bottles, squeeze fruit juice and sweep the floor, all of which the native did simultaneously. "Nonsense," Harvey croaked uncertainly. "We have seen enough queer things to know there are always more." He led the way inside. Through thirst-cracked lips he rasped: "Water—quick!" Without a word, the bartender reached under the counter, brought out two glasses of water. The interplanetary con-men drank noisily, asked for more, until they had drunk eight glasses. Meanwhile, the bartender had taken out eight jiggers and filled them with whiskey. Harvey and Joe were breathing hard from having gulped the water so fast, but they were beginning to revive. They noticed the bartender's impersonal eyes studying them shrewdly. "Strangers, eh?" he asked at last. "Solar salesmen, my colonial friend," Harvey answered in his usual lush manner. "We purvey that renowned Martian remedy, La-anago Yergis , the formula for which was recently discovered by ourselves in the ancient ruined city of La-anago. Medical science is unanimous in proclaiming this magic medicine the sole panacea in the entire history of therapeutics." "Yeah?" said the bartender disinterestedly, polishing the chaser glasses without washing them. "Where you heading?" "Out of Mars for Ganymede. Our condenser broke down, and we've gone without water for five ghastly days." "Got a mechanic around this dumping ground you call a port?" Joe asked. "We did. He came near starving and moved on to Titan. Ships don't land here unless they're in trouble." "Then where's the water lead-in? We'll fill up and push off." "Mayor takes care of that," replied the saloon owner. "If you gents're finished at the bar, your drinks'll be forty buckos." Harvey grinned puzzledly. "We didn't take any whiskey." "Might as well. Water's five buckos a glass. Liquor's free with every chaser." Harvey's eyes bulged. Joe gulped. "That—that's robbery!" the lanky man managed to get out in a thin quaver. The barkeeper shrugged. "When there ain't many customers, you gotta make more on each one. Besides—" "Besides nothing!" Joe roared, finding his voice again. "You dirty crook—robbing poor spacemen! You—" "You dirty crook!" Joe roared. "Robbing honest spacemen!" Harvey nudged him warningly. "Easy, my boy, easy." He turned to the bartender apologetically. "Don't mind my friend. His adrenal glands are sometimes overactive. You were going to say—?" The round face of the barkeeper had assumed an aggrieved expression. "Folks are always thinkin' the other feller's out to do 'em," he said, shaking his head. "Lemme explain about the water here. It's bitter as some kinds of sin before it's purified. Have to bring it in with buckets and make it sweet. That takes time and labor. Waddya think—I was chargin' feller critters for water just out of devilment? I charge because I gotta." "Friend," said Harvey, taking out a wallet and counting off eight five-bucko bills, "here is your money. What's fair is fair, and you have put a different complexion on what seemed at first to be an unconscionable interjection of a middleman between Nature and man's thirst." The saloon man removed his dirty apron and came around the bar. "If that's an apology, I accept it. Now the mayor'll discuss filling your tanks. That's me. I'm also justice of the peace, official recorder, fire chief...." "And chief of police, no doubt," said Harvey jocosely. "Nope. That's my son, Jed. Angus Johnson's my name. Folks here just call me Chief. I run this town, and run it right. How much water will you need?" Joe estimated quickly. "About seventy-five liters, if we go on half rations," he answered. He waited apprehensively. "Let's say ten buckos a liter," the mayor said. "On account of the quantity, I'm able to quote a bargain price. Shucks, boys, it hurts me more to charge for water than it does for you to pay. I just got to, that's all." The mayor gestured to the native, who shuffled out to the tanks with them. The planetoid man worked the pump while the mayor intently watched the crude level-gauge, crying "Stop!" when it registered the proper amount. Then Johnson rubbed his thumb on his index finger and wetted his lips expectantly. Harvey bravely counted off the bills. He asked: "But what are we to do about replenishing our battery fluid? Ten buckos a liter would be preposterous. We simply can't afford it." Johnson's response almost floored them. "Who said anything about charging you for battery water? You can have all you want for nothing. It's just the purified stuff that comes so high." After giving them directions that would take them to the free-water pool, the ponderous factotum of Planetoid 42 shook hands and headed back to the saloon. His six-armed assistant followed him inside. "Now do you see, my hot-tempered colleague?" said Harvey as he and Joe picked up buckets that hung on the tank. "Johnson, as I saw instantly, is the victim of a difficult environment, and must charge accordingly." "Just the same," Joe griped, "paying for water isn't something you can get used to in ten minutes." In the fragile forest, they soon came across a stream that sprang from the igneous soil and splashed into the small pond whose contents, according to the mayor, was theirs for the asking. They filled their buckets and hauled them to the ship, then returned for more. It was on the sixth trip that Joe caught a glimpse of Jupiter-shine on a bright surface off to the left. The figure, 750, with the bucko sign in front of it, was still doing acrobatics inside his skull and keeping a faint suspicion alive in him. So he called Harvey and they went to investigate. Among the skimpy ground-crawling vines, they saw a long slender mound that was unmistakably a buried pipe. "What's this doing here?" Harvey asked, puzzled. "I thought Johnson had to transport water in pails." "Wonder where it leads to," Joe said uneasily. "It leads to the saloon," said Harvey, his eyes rapidly tracing the pipe back toward the spaceport. "What I am concerned with is where it leads from ." Five minutes later, panting heavily from the unaccustomed exertion of scrambling through the tangle of planetorial undergrowth, they burst into the open—before a clear, sparkling pool. Mutely, Harvey pointed out a pipe-end jutting under the water. "I am growing suspicious," he said in a rigidly controlled voice. But Joe was already on his knees, scooping up a handful of water and tasting it. "Sweet!" he snarled. They rushed back to the first pool, where Joe again tasted a sample. His mouth went wry. "Bitter! He uses only one pool, the sweet one! The only thing that needs purifying around here is that blasted mayor's conscience." "The asteroidal Poobah has tricked us with a slick come-on," said Harvey slowly. His eyes grew cold. "Joseph, the good-natured artist in me has become a hard and merciless avenger. I shall not rest until we have had the best of this colonial con-man! Watch your cues from this point hence." Fists clenched, the two returned to the saloon. But at the door they stopped and their fists unclenched. "Thought you gents were leaving," the mayor called out, seeing them frozen in the doorway. "Glad you didn't. Now you can meet my son, Jed. Him and me are the whole Earthman population of Johnson City." "You don't need any more," said Harvey, dismayed. Johnson's eight-foot son, topped by a massive roof of sun-bleached hair and held up by a foundation that seemed immovable, had obviously been born and raised in low gravity. For any decent-sized world would have kept him down near the general dimensions of a man. He held out an acre of palm. Harvey studied it worriedly, put his own hand somewhere on it, swallowed as it closed, then breathed again when his fingers were released in five units instead of a single compressed one. "Pleased to meet you," piped a voice that had never known a dense atmosphere. The pursuit of vengeance, Harvey realized, had taken a quick and unpleasant turn. Something shrewd was called for.... "Joseph!" he exclaimed, looking at his partner in alarm. "Don't you feel well?" Even before the others could turn to him, Joe's practiced eyes were gently crossing. He sagged against the door frame, all his features drooping like a bloodhound's. "Bring him in here!" Johnson cried. "I mean, get him away! He's coming down with asteroid fever!" "Of course," replied Harvey calmly. "Any fool knows the first symptoms of the disease that once scourged the universe." "What do you mean, once ?" demanded Johnson. "I come down with it every year, and I ain't hankering to have it in an off-season. Get him out of here!" "In good time. He can't be moved immediately." "Then he'll be here for months!" Harvey helped Joe to the counter and lifted him up on it. The mayor and his gigantic offspring were cowering across the room, trying to breathe in tiny, uncontaminating gasps. "You'll find everything you want in the back room," Johnson said frantically, "sulfopyridine, mustard plasters, rubs, inhalers, suction cups—" "Relics of the past," Harvey stated. "One medication is all modern man requires to combat the dread menace, asteroid fever." "What's that?" asked the mayor without conviction. Instead of replying, Harvey hurried outside to the ungainly second-hand rocket ship in the center of the shabby spaceport. He returned within a few minutes, carrying a bottle. Joe was still stretched out on the bar, panting, his eyes slowly crossing and uncrossing. Harvey lifted the patient's head tenderly, put the bottle to his lips and tilted it until he was forced to drink. When Joe tried to pull away, Harvey was inexorable. He made his partner drink until most of the liquid was gone. Then he stepped back and waited for the inevitable result. Joe's performance was better than ever. He lay supine for several moments, his face twisted into an expression that seemed doomed to perpetual wryness. Slowly, however, he sat up and his features straightened out. "Are—are you all right?" asked the mayor anxiously. "Much better," said Joe in a weak voice. "Maybe you need another dose," Harvey suggested. Joe recoiled. "I'm fine now!" he cried, and sprang off the bar to prove it. Astonished, Johnson and his son drew closer. They searched Joe's face, and then the mayor timidly felt his pulse. "Well, I'll be hanged!" Johnson ejaculated. " La-anago Yergis never fails, my friend," Harvey explained. "By actual test, it conquers asteroid fever in from four to twenty-three minutes, depending on the severity of the attack. Luckily, we caught this one before it grew formidable." The mayor's eyes became clouded mirrors of an inward conflict. "If you don't charge too much," he said warily, "I might think of buying some." "We do not sell this unbelievable remedy," Harvey replied with dignity. "It sells itself." "'Course, I'd expect a considerable reduction if I bought a whole case," said Johnson. "That would be the smallest investment you could make, compared with the vast loss of time and strength the fever involves." "How much?" asked the mayor unhappily. "For you, since you have taken us in so hospitably, a mere five hundred buckos." Johnson did not actually stagger back, but he gave the impression of doing so. "F-four hundred," he offered. "Not a red cent less than four seventy-five," Harvey said flatly. "Make it four fifty," quavered Johnson. "I dislike haggling," said Harvey. The final price, however, was four hundred and sixty-nine buckos and fifty redsents. Magnanimously, Harvey added: "And we will include, gratis , an elegant bottle-opener, a superb product of Mercurian handicraftsmanship." Johnson stabbed out a warning finger. "No tricks now. I want a taste of that stuff. You're not switching some worthless junk on me." Harvey took a glass from the bar and poured him a generous sample. The mayor sniffed it, grimaced, then threw it down his gullet. The ensuing minute saw a grim battle between a man and his stomach, a battle which the man gradually won. "There ain't no words for that taste," he gulped when it was safe to talk again. "Medicine," Harvey propounded, "should taste like medicine." To Joe he said: "Come, my esteemed colleague. We must perform the sacred task to which we have dedicated ourselves." With Joe stumbling along behind, he left the saloon, crossed the clearing and entered the ship. As soon as they were inside, Joe dropped his murderous silence and cried: "What kind of a dirty trick was that, giving me poison instead of that snake oil?" "That was not poison," Harvey contradicted quietly. "It was La-anago Yergis extract, plus." "Plus what—arsenic?" "Now, Joseph! Consider my quandary when I came back here to manufacture our specific for all known ailments, with the intention of selling yonder asteroidal tin-horn a bill of medical goods—an entire case, mind you. Was I to mix the extract with the water for which we had been swindled to the tune of ten buckos a liter? Where would our profit have been, then? No; I had to use the bitter free water, of course." "But why use it on me?" Joe demanded furiously. Harvey looked reprovingly at his gangling partner. "Did Johnson ask to taste it, or did he not? One must look ahead, Joseph. I had to produce the same medicine that we will now manufacture. Thus, you were a guinea pig for a splendid cause." "Okay, okay," Joe said. "But you shoulda charged him more." "Joseph, I promise you that we shall get back every redsent of which that swindler cheated us, besides whatever other funds or valuables he possesses. We could not be content with less." "Well, we're starting all right," admitted Joe. "How about that thing with six arms? He looks like a valuable. Can't we grab him off?" Harvey stopped filling bottles and looked up pensively. "I have every hope of luring away the profitable monstrosity. Apparently you have also surmised the fortune we could make with him. At first I purpose to exhibit him on our interplanetary tours with our streamlined panacea; he would be a spectacular attraction for bucolic suckers. Later, a brief period of demonstrating his abilities on the audio-visiphone. Then our triumph—we shall sell him at a stupendous figure to the zoo!" Joe was still dazed by that monetary vista when he and Harvey carried the case of medicine to the saloon. The mayor had already cleared a place of honor in the cluttered back room, where he told them to put it down carefully. Then he took the elaborate bottle-opener Harvey gave him, reverently uncorked a bottle and sampled it. It must have been at least as good as the first; he gagged. "That's the stuff, all right," he said, swallowing hard. He counted out the money into Harvey's hand, at a moderate rate that precariously balanced between his pleasure at getting the fever remedy and his pain at paying for it. Then he glanced out to see the position of Jupiter, and asked: "You gents eaten yet? The restaurant's open now." Harvey and Joe looked at each other. They hadn't been thinking about food at all, but suddenly they realized that they were hungry. "It's only water we were short of," Harvey said apprehensively. "We've got rations back at the ship." " H-mph! " the mayor grunted. "Powdered concentrates. Compressed pap. Suit yourselves. We treat our stomachs better here. And you're welcome to our hospitality." "Your hospitality," said Harvey, "depends on the prices you charge." "Well, if that's what's worrying you, you can stop worrying," answered the mayor promptly. "What's more, the kind of dinner I serve here you can't get anywhere else for any price." Swiftly, Harvey conned the possibilities of being bilked again. He saw none. "Let's take a look at the menu, anyhow, Joe," he said guardedly. Johnson immediately fell into the role of "mine host." "Come right in, gents," he invited. "Right into the dining room." He seated them at a table, which a rope tied between posts made more or less private, though nobody else was in the saloon and there was little chance of company. Genius, the six-armed native, appeared from the dingy kitchen with two menus in one hand, two glasses of water in another, plus napkins, silverware, a pitcher, plates, saucers, cups, and their cocktails, which were on the house. Then he stood by for orders. Harvey and Joe studied the menu critically. The prices were phenomenally low. When they glanced up at Johnson in perplexity, he grinned, bowed and asked: "Everything satisfactory, gents?" "Quite," said Harvey. "We shall order." For an hour they were served amazing dishes, both fresh and canned, the culinary wealth of this planetoid and all the system. And the service was as extraordinary as the meal itself. With four hands, Genius played deftly upon a pair of mellow Venusian viotars , using his other two hands for waiting on the table. "We absolutely must purchase this incredible specimen," Harvey whispered excitedly when Johnson and the native were both in the kitchen, attending to the next course. "He would make any society hostess's season a riotous success, which should be worth a great sum to women like Mrs. van Schuyler-Morgan, merely for his hire." "Think of a fast one fast," Joe agreed. "You're right." "But I dislike having to revise my opinion of a man so often," complained Harvey. "I wish Johnson would stay either swindler or honest merchant. This dinner is worth as least twenty buckos, yet I estimate our check at a mere bucko twenty redsents." The mayor's appearance prevented them from continuing the discussion. "It's been a great honor, gents," he said. "Ain't often I have visitors, and I like the best, like you two gents." As if on cue, Genius came out and put the check down between Joe and Harvey. Harvey picked it up negligently, but his casual air vanished in a yelp of horror. "What the devil is this?" he shouted.—"How do you arrive at this fantastic, idiotic figure— three hundred and twenty-eight buckos !" Johnson didn't answer. Neither did Genius; he simply put on the table, not a fingerbowl, but a magnifying glass. With one of his thirty fingers he pointed politely to the bottom of the menu. Harvey focused on the microscopic print, and his face went pasty with rage. The minute note read: "Services and entertainment, 327 buckos 80 redsents." "You can go to hell!" Joe growled. "We won't pay it!" Johnson sighed ponderously. "I was afraid you'd act like that," he said with regret. He pulled a tin badge out of his rear pocket, pinned it on his vest, and twisted his holstered gun into view. "Afraid I'll have to ask the sheriff to take over." Johnson, the "sheriff," collected the money, and Johnson, the "restaurateur," pocketed it. Meanwhile, Harvey tipped Joe the sign to remain calm. "My friend," he said to the mayor, and his tones took on a schoolmasterish severity, "your long absence from Earth has perhaps made you forget those elements of human wisdom that have entered the folk-lore of your native planet. Such as, for example: 'It is folly to kill a goose that lays golden eggs,' and 'Penny wise is pound foolish.'" "I don't get the connection," objected Johnson. "Well, by obliging us to pay such a high price for your dinner, you put out of your reach the chance of profiting from a really substantial deal. My partner and I were prepared to make you a sizable offer for the peculiar creature you call Genius. But by reducing our funds the way you have—" "Who said I wanted to sell him?" the mayor interrupted. He rubbed his fingers together and asked disinterestedly: "What were you going to offer, anyhow?" "It doesn't matter any longer," Harvey said with elaborate carelessness. "Perhaps you wouldn't have accepted it, anyway." "That's right," Johnson came back emphatically. "But what would your offer have been which I would have turned down?" "Which one? The one we were going to make, or the one we can make now?" "Either one. It don't make no difference. Genius is too valuable to sell." "Oh, come now, Mr. Johnson. Don't tell me no amount of money would tempt you!" "Nope. But how much did you say?" "Ah, then you will consider releasing Genius!" "Well, I'll tell you something," said the mayor confidentially. "When you've got one thing, you've got one thing. But when you've got money, it's the same as having a lot of things. Because, if you've got money, you can buy this and that and this and that and—" "This and that," concluded Joe. "We'll give you five hundred buckos." "Now, gents!" Johnson remonstrated. "Why, six hundred would hardly—" "You haven't left us much money," Harvey put in. The mayor frowned. "All right, we'll split the difference. Make it five-fifty." Harvey was quick to pay out, for this was a genuine windfall. Then he stood up and admired the astonishing possession he had so inexpensively acquired. "I really hate to deprive you of this unique creature," he said to Johnson. "I should imagine you will be rather lonely, with only your filial mammoth to keep you company." "I sure will," Johnson confessed glumly. "I got pretty attached to Genius, and I'm going to miss him something awful." Harvey forcibly removed his eyes from the native, who was clearing off the table almost all at once. "My friend," he said, "we take your only solace, it is true, but in his place we can offer something no less amazing and instructive." The mayor's hand went protectively to his pocket. "What is it?" he asked with the suspicion of a man who has seen human nature at its worst and expects nothing better. "Joseph, get our most prized belonging from the communications room of the ship," Harvey instructed. To Johnson he explained: "You must see the wondrous instrument before its value can be appreciated. My partner will soon have it here for your astonishment." Joe's face grew as glum as Johnson's had been. "Aw, Harv," he protested, "do we have to sell it? And right when I thought we were getting the key!" "We must not be selfish, my boy," Harvey said nobly. "We have had our chance; now we must relinquish Fate to the hands of a man who might have more success than we. Go, Joseph. Bring it here." Unwillingly, Joe turned and shuffled out. On a larger and heavier world than Planetoid 42, Johnson's curiosity would probably have had weight and mass. He was bursting with questions, but he was obviously afraid they would cost him money. For his part, Harvey allowed that curiosity to grow like a Venusian amoeba until Joe came in, lugging a radio. "Is that what you were talking about?" the mayor snorted. "What makes you think I want a radio? I came here to get away from singers and political speech-makers." "Do not jump to hasty conclusions," Harvey cautioned. "Another word, and I shall refuse you the greatest opportunity any man has ever had, with the sole exceptions of Joseph, myself and the unfortunate inventor of this absolutely awe-inspiring device." "I ain't in the market for a radio," Johnson said stubbornly. Harvey nodded in relief. "We have attempted to repay our host, Joseph. He has spurned our generosity. We have now the chance to continue our study, which I am positive will soon reward us with the key to an enormous fortune." "Well, that's no plating off our bow," Joe grunted. "I'm glad he did turn it down. I hated to give it up after working on it for three whole years." He picked up the radio and began walking toward the door. "Now, hold on!" the mayor cried. "I ain't saying I'll buy, but what is it I'm turning down?" Joe returned and set the instrument down on the bar. His face sorrowful, Harvey fondly stroked the scarred plasticoid cabinet. "To make a long story, Mr. Johnson," he said, "Joseph and I were among the chosen few who knew the famous Doctor Dean intimately. Just before his tragic death, you will recall, Dean allegedly went insane." He banged his fist on the bar. "I have said it before, and I repeat again, that was a malicious lie, spread by the doctor's enemies to discredit his greatest invention—this fourth dimensional radio!" "This what?" Johnson blurted out. "In simple terms," clarified Harvey, "the ingenious doctor discovered that the yawning chasm between the dimensions could be bridged by energy of all quanta. There has never been any question that the inhabitants of the super-dimension would be far more civilized than ourselves. Consequently, the man who could tap their knowledge would find himself in possession of a powerful, undreamt-of science!" The mayor looked respectfully at the silent box on the bar. "And this thing gets broadcasts from the fourth dimension?" "It does, Mr. Johnson! Only charlatans like those who envied Doctor Dean's magnificent accomplishments could deny that fact." The mayor put his hands in his pockets, unswiveled one hip and stared thoughtfully at the battered cabinet. "Well, let's say it picks up fourth dimensional broadcasts," he conceded. "But how could you understand what they're saying? Folks up there wouldn't talk our language." Again Harvey smashed his fist down. "Do you dare to repeat the scurvy lie that broke Dean's spirit and drove him to suicide?" Johnson recoiled. "No—no, of course not . I mean, being up here, I naturally couldn't get all the details." "Naturally," Harvey agreed, mollified. "I'm sorry I lost my temper. But it is a matter of record that the doctor proved the broadcasts emanating from the super-dimension were in English! Why should that be so difficult to believe? Is it impossible that at one time there was communication between the dimensions, that the super-beings admired our language and adopted it in all its beauty, adding to it their own hyper-scientific trimmings?" "Why, I don't know," Johnson said in confusion. "For three years, Joseph and I lost sleep and hair, trying to detect the simple key that would translate the somewhat metamorphosed broadcasts into our primitive English. It eluded us. Even the doctor failed. But that was understandable; a sensitive soul like his could stand only so much. And the combination of ridicule and failure to solve the mystery caused him to take his own life." Johnson winced. "Is that what you want to unload on me?" "For a very good reason, sir. Patience is the virtue that will be rewarded with the key to these fourth dimensional broadcasts. A man who could devote his life to improving this lonely worldlet is obviously a person with unusual patience." "Yeah," the mayor said grudgingly, "I ain't exactly flighty." "Therefore, you are the man who could unravel the problem!" Johnson asked skeptically: "How about a sample first?" | B. Joe's acting skills |
What the system designs introduced? | ### Introduction
Amazon Alexa Prize BIBREF0 provides a platform to collect real human-machine conversation data and evaluate performance on speech-based social conversational systems. Our system, Gunrock BIBREF1 addresses several limitations of prior chatbots BIBREF2, BIBREF3, BIBREF4 including inconsistency and difficulty in complex sentence understanding (e.g., long utterances) and provides several contributions: First, Gunrock's multi-step language understanding modules enable the system to provide more useful information to the dialog manager, including a novel dialog act scheme. Additionally, the natural language understanding (NLU) module can handle more complex sentences, including those with coreference. Second, Gunrock interleaves actions to elicit users' opinions and provide responses to create an in-depth, engaging conversation; while a related strategy to interleave task- and non-task functions in chatbots has been proposed BIBREF5, no chatbots to our knowledge have employed a fact/opinion interleaving strategy. Finally, we use an extensive persona database to provide coherent profile information, a critical challenge in building social chatbots BIBREF3. Compared to previous systems BIBREF4, Gunrock generates more balanced conversations between human and machine by encouraging and understanding more human inputs (see Table TABREF2 for an example). ### System Architecture
Figure FIGREF3 provides an overview of Gunrock's architecture. We extend the Amazon Conversational Bot Toolkit (CoBot) BIBREF6 which is a flexible event-driven framework. CoBot provides ASR results and natural language processing pipelines through the Alexa Skills Kit (ASK) BIBREF7. Gunrock corrects ASR according to the context (asr) and creates a natural language understanding (NLU) (nlu) module where multiple components analyze the user utterances. A dialog manager (DM) (dm) uses features from NLU to select topic dialog modules and defines an individual dialog flow. Each dialog module leverages several knowledge bases (knowledge). Then a natural language generation (NLG) (nlg) module generates a corresponding response. Finally, we markup the synthesized responses and return to the users through text to speech (TTS) (tts). While we provide an overview of the system in the following sections, for detailed system implementation details, please see the technical report BIBREF1. ### System Architecture ::: Automatic Speech Recognition
Gunrock receives ASR results with the raw text and timestep information for each word in the sequence (without case information and punctuation). Keywords, especially named entities such as movie names, are prone to generate ASR errors without contextual information, but are essential for NLU and NLG. Therefore, Gunrock uses domain knowledge to correct these errors by comparing noun phrases to a knowledge base (e.g. a list of the most popular movies names) based on their phonetic information. We extract the primary and secondary code using The Double Metaphone Search Algorithm BIBREF8 for noun phrases (extracted by noun trunks) and the selected knowledge base, and suggest a potential fix by code matching. An example can be seen in User_3 and Gunrock_3 in Table TABREF2. ### System Architecture ::: Natural Language Understanding
Gunrock is designed to engage users in deeper conversation; accordingly, a user utterance can consist of multiple units with complete semantic meanings. We first split the corrected raw ASR text into sentences by inserting break tokens. An example is shown in User_3 in Table TABREF2. Meanwhile, we mask named entities before segmentation so that a named entity will not be segmented into multiple parts and an utterance with a complete meaning is maintained (e.g.,“i like the movie a star is born"). We also leverage timestep information to filter out false positive corrections. After segmentation, our coreference implementation leverages entity knowledge (such as person versus event) and replaces nouns with their actual reference by entity ranking. We implement coreference resolution on entities both within segments in a single turn as well as across multiple turns. For instance, “him" in the last segment in User_5 is replaced with “bradley cooper" in Table TABREF2. Next, we use a constituency parser to generate noun phrases from each modified segment. Within the sequence pipeline to generate complete segments, Gunrock detects (1) topic, (2) named entities, and (3) sentiment using ASK in parallel. The NLU module uses knowledge graphs including Google Knowledge Graph to call for a detailed description of each noun phrase for understanding. In order to extract the intent for each segment, we designed MIDAS, a human-machine dialog act scheme with 23 tags and implemented a multi-label dialog act classification model using contextual information BIBREF9. Next, the NLU components analyzed on each segment in a user utterance are sent to the DM and NLG module for state tracking and generation, respectively. ### System Architecture ::: Dialog Manager
We implemented a hierarchical dialog manager, consisting of a high level and low level DMs. The former leverages NLU outputs for each segment and selects the most important segment for the system as the central element using heuristics. For example, “i just finished reading harry potter," triggers Sub-DM: Books. Utilizing the central element and features extracted from NLU, input utterances are mapped onto 11 possible topic dialog modules (e.g., movies, books, animals, etc.), including a backup module, retrieval. Low level dialog management is handled by the separate topic dialog modules, which use modular finite state transducers to execute various dialog segments processed by the NLU. Using topic-specific modules enables deeper conversations that maintain the context. We design dialog flows in each of the finite state machines, as well. Dialog flow is determined by rule-based transitions between a specified fixed set of dialog states. To ensure that our states and transitions are effective, we leverage large scale user data to find high probability responses and high priority responses to handle in different contexts. Meanwhile, dialog flow is customized to each user by tracking user attributes as dialog context. In addition, each dialog flow is adaptive to user responses to show acknowledgement and understanding (e.g., talking about pet ownership in the animal module). Based on the user responses, many dialog flow variations exist to provide a fresh experience each time. This reduces the feeling of dialogs being scripted and repetitive. Our dialog flows additionally interleave facts, opinions, experiences, and questions to make the conversation flexible and interesting. In the meantime, we consider feedback signals such as “continue" and “stop" from the current topic dialog module, indicating whether it is able to respond to the following request in the dialog flow, in order to select the best response module. Additionally, in all modules we allow mixed-initiative interactions; users can trigger a new dialog module when they want to switch topics while in any state. For example, users can start a new conversation about movies from any other topic module. ### System Architecture ::: Knowledge Databases
All topic dialog modules query knowledge bases to provide information to the user. To respond to general factual questions, Gunrock queries the EVI factual database , as well as other up-to-date scraped information appropriate for the submodule, such as news and current showing movies in a specific location from databases including IMDB. One contribution of Gunrock is the extensive Gunrock Persona Backstory database, consisting of over 1,000 responses to possible questions for Gunrock as well as reasoning for her responses for roughly 250 questions (see Table 2). We designed the system responses to elicit a consistent personality within and across modules, modeled as a female individual who is positive, outgoing, and is interested in science and technology. ### System Architecture ::: Natural Language Generation
In order to avoid repetitive and non-specific responses commonly seen in dialog systems BIBREF10, Gunrock uses a template manager to select from a handcrafted response templates based on the dialog state. One dialog state can map to multiple response templates with similar semantic or functional content but differing surface forms. Among these response templates for the same dialog state, one is randomly selected without repetition to provide variety unless all have been exhausted. When a response template is selected, any slots are substituted with actual contents, including queried information for news and specific data for weather. For example, to ground a movie name due to ASR errors or multiple versions, one template is “Are you talking about {movie_title} released in {release_year} starring {actor_name} as {actor_role}?". Module-specific templates were generated for each topic (e.g., animals), but some of the templates are generalizable across different modules (e.g., “What’s your favorite [movie $|$ book $|$ place to visit]?") In many cases, response templates corresponding to different dialog acts are dynamically composed to give the final response. For example, an appropriate acknowledgement for the user’s response can be combined with a predetermined follow-up question. ### System Architecture ::: Text To Speech
After NLG, we adjust the TTS of the system to improve the expressiveness of the voice to convey that the system is an engaged and active participant in the conversation. We use a rule-based system to systematically add interjections, specifically Alexa Speechcons, and fillers to approximate human-like cognitive-emotional expression BIBREF11. For more on the framework and analysis of the TTS modifications, see BIBREF12. ### Analysis
From January 5, 2019 to March 5, 2019, we collected conversational data for Gunrock. During this time, no other code updates occurred. We analyzed conversations for Gunrock with at least 3 user turns to avoid conversations triggered by accident. Overall, this resulted in a total of 34,432 user conversations. Together, these users gave Gunrock an average rating of 3.65 (median: 4.0), which was elicited at the end of the conversation (“On a scale from 1 to 5 stars, how do you feel about talking to this socialbot again?"). Users engaged with Gunrock for an average of 20.92 overall turns (median 13.0), with an average of 6.98 words per utterance, and had an average conversation time of 7.33 minutes (median: 2.87 min.). We conducted three principal analyses: users' response depth (wordcount), backstory queries (backstorypersona), and interleaving of personal and factual responses (pets). ### Analysis ::: Response Depth: Mean Word Count
Two unique features of Gunrock are its ability to dissect longer, complex sentences, and its methods to encourage users to be active conversationalists, elaborating on their responses. In prior work, even if users are able to drive the conversation, often bots use simple yes/no questions to control the conversational flow to improve understanding; as a result, users are more passive interlocutors in the conversation. We aimed to improve user engagement by designing the conversation to have more open-ended opinion/personal questions, and show that the system can understand the users' complex utterances (See nlu for details on NLU). Accordingly, we ask if users' speech behavior will reflect Gunrock's technical capability and conversational strategy, producing longer sentences. We assessed the degree of conversational depth by measuring users' mean word count. Prior work has found that an increase in word count has been linked to improved user engagement (e.g., in a social dialog system BIBREF13). For each user conversation, we extracted the overall rating, the number of turns of the interaction, and the user's per-utterance word count (averaged across all utterances). We modeled the relationship between word count and the two metrics of user engagement (overall rating, mean number of turns) in separate linear regressions. Results showed that users who, on average, produced utterances with more words gave significantly higher ratings ($\beta $=0.01, SE=0.002, t=4.79, p$<$0.001)(see Figure 2) and engaged with Gunrock for significantly greater number of turns ($\beta $=1.85, SE=0.05, t=35.58, p$<$0.001) (see Figure 2). These results can be interpreted as evidence for Gunrock's ability to handle complex sentences, where users are not constrained to simple responses to be understood and feel engaged in the conversation – and evidence that individuals are more satisfied with the conversation when they take a more active role, rather than the system dominating the dialog. On the other hand, another interpretation is that users who are more talkative may enjoy talking to the bot in general, and thus give higher ratings in tandem with higher average word counts. ### Analysis ::: Gunrock's Backstory and Persona
We assessed the user's interest in Gunrock by tagging instances where the user triggered Gunrock's backstory (e.g., “What's your favorite color?"). For users with at least one backstory question, we modeled overall (log) Rating with a linear regression by the (log) `Number of Backstory Questions Asked' (log transformed due to the variables' nonlinear relationship). We hypothesized that users who show greater curiosity about Gunrock will display higher overall ratings for the conversation based on her responses. Overall, the number of times users queried Gunrock's backstory was strongly related to the rating they gave at the end of the interaction (log:$\beta $=0.10, SE=0.002, t=58.4, p$<$0.001)(see Figure 3). This suggests that maintaining a consistent personality — and having enough responses to questions the users are interested in — may improve user satisfaction. ### Analysis ::: Interleaving Personal and Factual Information: Animal Module
Gunrock includes a specific topic module on animals, which includes a factual component where the system provides animal facts, as well as a more personalized component about pets. Our system is designed to engage users about animals in a more casual conversational style BIBREF14, eliciting follow-up questions if the user indicates they have a pet; if we are able to extract the pet's name, we refer to it in the conversation (e.g., “Oliver is a great name for a cat!", “How long have you had Oliver?"). In cases where the user does not indicate that they have a pet, the system solely provides animal facts. Therefore, the animal module can serve as a test of our interleaving strategy: we hypothesized that combining facts and personal questions — in this case about the user's pet — would lead to greater user satisfaction overall. We extracted conversations where Gunrock asked the user if they had ever had a pet and categorized responses as “Yes", “No", or “NA" (if users did not respond with an affirmative or negative response). We modeled user rating with a linear regression model, with predictor of “Has Pet' (2 levels: Yes, No). We found that users who talked to Gunrock about their pet showed significantly higher overall ratings of the conversation ($\beta $=0.15, SE=0.06, t=2.53, p$=$0.016) (see Figure 4). One interpretation is that interleaving factual information with more in-depth questions about their pet result in improved user experience. Yet, another interpretation is that pet owners may be more friendly and amenable to a socialbot; for example, prior research has linked differences in personality to pet ownership BIBREF15. ### Conclusion
Gunrock is a social chatbot that focuses on having long and engaging speech-based conversations with thousands of real users. Accordingly, our architecture employs specific modules to handle longer and complex utterances and encourages users to be more active in a conversation. Analysis shows that users' speech behavior reflects these capabilities. Longer sentences and more questions about Gunrocks's backstory positively correlate with user experience. Additionally, we find evidence for interleaved dialog flow, where combining factual information with personal opinions and stories improve user satisfaction. Overall, this work has practical applications, in applying these design principles to other social chatbots, as well as theoretical implications, in terms of the nature of human-computer interaction (cf. 'Computers are Social Actors' BIBREF16). Our results suggest that users are engaging with Gunrock in similar ways to other humans: in chitchat about general topics (e.g., animals, movies, etc.), taking interest in Gunrock's backstory and persona, and even producing more information about themselves in return. ### Acknowledgments
We would like to acknowledge the help from Amazon in terms of financial and technical support. Figure 1: Gunrock system architecture Figure 2: Mean user rating by mean number of words. Error bars show standard error. Figure 3: Mean user rating based on number of queries to Gunrock’s backstory. Error bars show standard error. Figure 4: Mean user rating based ’Has Pet’. Error bars show standard error. | Amazon Conversational Bot Toolkit, natural language understanding (NLU) (nlu) module, dialog manager, knowledge bases, natural language generation (NLG) (nlg) module, text to speech (TTS) (tts) |
Why did the recruiter offer Lawrence $50 Galactic a week?
A. That was what was promised to all travelers to Earth for display.
B. He was able to offer him less, knowing he would still accept and be grateful.
C. He could be paid less because he was smaller and less of an attraction.
D. He would be paid less because he would also be reimbursed for expenses and have free travel.
| Birds of a Feather By ROBERT SILVERBERG Illustrated by WOOD [Transcriber's Note: This etext was produced from Galaxy Magazine November 1958. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed.] Getting specimens for the interstellar zoo was no problem—they battled for the honor—but now I had to fight like a wildcat to keep a display from making a monkey of me! It was our first day of recruiting on the planet, and the alien life-forms had lined up for hundreds of feet back from my rented office. As I came down the block from the hotel, I could hear and see and smell them with ease. My three staff men, Auchinleck, Stebbins and Ludlow, walked shieldwise in front of me. I peered between them to size the crop up. The aliens came in every shape and form, in all colors and textures—and all of them eager for a Corrigan contract. The Galaxy is full of bizarre beings, but there's barely a species anywhere that can resist the old exhibitionist urge. "Send them in one at a time," I told Stebbins. I ducked into the office, took my place back of the desk and waited for the procession to begin. The name of the planet was MacTavish IV (if you went by the official Terran listing) or Ghryne (if you called it by what its people were accustomed to calling it). I thought of it privately as MacTavish IV and referred to it publicly as Ghryne. I believe in keeping the locals happy wherever I go. Through the front window of the office, I could see our big gay tridim sign plastered to a facing wall: WANTED—EXTRATERRESTRIALS! We had saturated MacTavish IV with our promotional poop for a month preceding arrival. Stuff like this: Want to visit Earth—see the Galaxy's most glittering and exclusive world? Want to draw good pay, work short hours, experience the thrills of show business on romantic Terra? If you are a non-terrestrial, there may be a place for you in the Corrigan Institute of Morphological Science. No freaks wanted—normal beings only. J. F. Corrigan will hold interviews in person on Ghryne from Thirdday to Fifthday of Tenmonth. His last visit to the Caledonia Cluster until 2937, so don't miss your chance! Hurry! A life of wonder and riches can be yours! Broadsides like that, distributed wholesale in half a thousand languages, always bring them running. And the Corrigan Institute really packs in the crowds back on Earth. Why not? It's the best of its kind, the only really decent place where Earthmen can get a gander at the other species of the universe. The office buzzer sounded. Auchinleck said unctuously, "The first applicant is ready to see you, sir." "Send him, her or it in." The door opened and a timid-looking life-form advanced toward me on nervous little legs. He was a globular creature about the size of a big basketball, yellowish-green, with two spindly double-kneed legs and five double-elbowed arms, the latter spaced regularly around his body. There was a lidless eye at the top of his head and five lidded ones, one above each arm. Plus a big, gaping, toothless mouth. His voice was a surprisingly resounding basso. "You are Mr. Corrigan?" "That's right." I reached for a data blank. "Before we begin, I'll need certain information about—" "I am a being of Regulus II," came the grave, booming reply, even before I had picked up the blank. "I need no special care and I am not a fugitive from the law of any world." "Your name?" "Lawrence R. Fitzgerald." I throttled my exclamation of surprise, concealing it behind a quick cough. "Let me have that again, please?" "Certainly. My name is Lawrence R. Fitzgerald. The 'R' stands for Raymond." "Of course, that's not the name you were born with." The being closed his eyes and toddled around in a 360-degree rotation, remaining in place. On his world, that gesture is the equivalent of an apologetic smile. "My Regulan name no longer matters. I am now and shall evermore be Lawrence R. Fitzgerald. I am a Terraphile, you see." The little Regulan was as good as hired. Only the formalities remained. "You understand our terms, Mr. Fitzgerald?" "I'll be placed on exhibition at your Institute on Earth. You'll pay for my services, transportation and expenses. I'll be required to remain on exhibit no more than one-third of each Terran sidereal day." "And the pay will be—ah—$50 Galactic a week, plus expenses and transportation." The spherical creature clapped his hands in joy, three hands clapping on one side, two on the other. "Wonderful! I will see Earth at last! I accept the terms!" I buzzed for Ludlow and gave him the fast signal that meant we were signing this alien up at half the usual pay, and Ludlow took him into the other office to sign him up. I grinned, pleased with myself. We needed a green Regulan in our show; the last one had quit four years ago. But just because we needed him didn't mean we had to be extravagant in hiring him. A Terraphile alien who goes to the extent of rechristening himself with a Terran monicker would work for nothing, or even pay us, just so long as we let him get to Earth. My conscience won't let me really exploit a being, but I don't believe in throwing money away, either. The next applicant was a beefy ursinoid from Aldebaran IX. Our outfit has all the ursinoids it needs or is likely to need in the next few decades, and so I got rid of him in a couple of minutes. He was followed by a roly-poly blue-skinned humanoid from Donovan's Planet, four feet high and five hundred pounds heavy. We already had a couple of his species in the show, but they made good crowd-pleasers, being so plump and cheerful. I passed him along to Auchinleck to sign at anything short of top rate. Next came a bedraggled Sirian spider who was more interested in a handout than a job. If there's any species we have a real over-supply of, it's those silver-colored spiders, but this seedy specimen gave it a try anyway. He got the gate in half a minute, and he didn't even get the handout he was angling for. I don't approve of begging. The flora of applicants was steady. Ghryne is in the heart of the Caledonia Cluster, where the interstellar crossroads meet. We had figured to pick up plenty of new exhibits here and we were right. It was the isolationism of the late 29th century that turned me into the successful proprietor of Corrigan's Institute, after some years as an impoverished carnival man in the Betelgeuse system. Back in 2903, the World Congress declared Terra off-bounds for non-terrestrial beings, as an offshoot of the Terra for Terrans movement. Before then, anyone could visit Earth. After the gate clanged down, a non-terrestrial could only get onto Sol III as a specimen in a scientific collection—in short, as an exhibit in a zoo. That's what the Corrigan Institute of Morphological Science really is, of course. A zoo. But we don't go out and hunt for our specimens; we advertise and they come flocking to us. Every alien wants to see Earth once in his lifetime, and there's only one way he can do it. We don't keep too big an inventory. At last count, we had 690 specimens before this trip, representing 298 different intelligent life-forms. My goal is at least one member of at least 500 different races. When I reach that, I'll sit back and let the competition catch up—if it can. After an hour of steady work that morning, we had signed eleven new specimens. At the same time, we had turned away a dozen ursinoids, fifty of the reptilian natives of Ghryne, seven Sirian spiders, and no less than nineteen chlorine-breathing Procyonites wearing gas masks. It was also my sad duty to nix a Vegan who was negotiating through a Ghrynian agent. A Vegan would be a top-flight attraction, being some 400 feet long and appropriately fearsome to the eye, but I didn't see how we could take one on. They're gentle and likable beings, but their upkeep runs into literally tons of fresh meat a day, and not just any old kind of meat either. So we had to do without the Vegan. "One more specimen before lunch," I told Stebbins, "to make it an even dozen." He looked at me queerly and nodded. A being entered. I took a long close look at the life-form when it came in, and after that I took another one. I wondered what kind of stunt was being pulled. So far as I could tell, the being was quite plainly nothing but an Earthman. He sat down facing me without being asked and crossed his legs. He was tall and extremely thin, with pale blue eyes and dirty-blond hair, and though he was clean and reasonably well dressed, he had a shabby look about him. He said, in level Terran accents, "I'm looking for a job with your outfit, Corrigan." "There's been a mistake. We're interested in non-terrestrials only." "I'm a non-terrestrial. My name is Ildwar Gorb, of the planet Wazzenazz XIII." I don't mind conning the public from time to time, but I draw the line at getting bilked myself. "Look, friend, I'm busy, and I'm not known for my sense of humor. Or my generosity." "I'm not panhandling. I'm looking for a job." "Then try elsewhere. Suppose you stop wasting my time, bud. You're as Earthborn as I am." "I've never been within a dozen parsecs of Earth," he said smoothly. "I happen to be a representative of the only Earthlike race that exists anywhere in the Galaxy but on Earth itself. Wazzenazz XIII is a small and little-known planet in the Crab Nebula. Through an evolutionary fluke, my race is identical with yours. Now, don't you want me in your circus?" "No. And it's not a circus. It's—" "A scientific institute. I stand corrected." There was something glib and appealing about this preposterous phony. I guess I recognized a kindred spirit or I would have tossed him out on his ear without another word. Instead I played along. "If you're from such a distant place, how come you speak English so well?" "I'm not speaking. I'm a telepath—not the kind that reads minds, just the kind that projects. I communicate in symbols that you translate back to colloquial speech." "Very clever, Mr. Gorb." I grinned at him and shook my head. "You spin a good yarn—but for my money, you're really Sam Jones or Phil Smith from Earth, stranded here and out of cash. You want a free trip back to Earth. No deal. The demand for beings from Wazzenazz XIII is pretty low these days. Zero, in fact. Good-by, Mr. Gorb." He pointed a finger squarely at me and said, "You're making a big mistake. I'm just what your outfit needs. A representative of a hitherto utterly unknown race identical to humanity in every respect! Look here, examine my teeth. Absolutely like human teeth! And—" I pulled away from his yawning mouth. "Good-by, Mr. Gorb," I repeated. "All I ask is a contract, Corrigan. It isn't much. I'll be a big attraction. I'll—" " Good-by, Mr. Gorb! " He glowered at me reproachfully for a moment, stood up and sauntered to the door. "I thought you were a man of acumen, Corrigan. Well, think it over. Maybe you'll regret your hastiness. I'll be back to give you another chance." He slammed the door and I let my grim expression relax into a smile. This was the best con switch yet—an Earthman posing as an alien to get a job! But I wasn't buying it, even if I could appreciate his cleverness intellectually. There's no such place as Wazzenazz XIII and there's only one human race in the Galaxy—on Earth. I was going to need some real good reason before I gave a down-and-out grifter a free ticket home. I didn't know it then, but before the day was out, I would have that reason. And, with it, plenty of trouble on my hands. The first harbinger of woe turned up after lunch in the person of a Kallerian. The Kallerian was the sixth applicant that afternoon. I had turned away three more ursinoids, hired a vegetable from Miazan, and said no to a scaly pseudo-armadillo from one of the Delta Worlds. Hardly had the 'dillo scuttled dejectedly out of my office when the Kallerian came striding in, not even waiting for Stebbins to admit him officially. He was big even for his kind—in the neighborhood of nine feet high, and getting on toward a ton. He planted himself firmly on his three stocky feet, extended his massive arms in a Kallerian greeting-gesture, and growled, "I am Vallo Heraal, Freeman of Kaller IV. You will sign me immediately to a contract." "Sit down, Freeman Heraal. I like to make my own decisions, thanks." "You will grant me a contract!" "Will you please sit down?" He said sulkily, "I will remain standing." "As you prefer." My desk has a few concealed features which are sometimes useful in dealing with belligerent or disappointed life-forms. My fingers roamed to the meshgun trigger, just in case of trouble. The Kallerian stood motionless before me. They're hairy creatures, and this one had a coarse, thick mat of blue fur completely covering his body. Two fierce eyes glimmered out through the otherwise dense blanket of fur. He was wearing the kilt, girdle and ceremonial blaster of his warlike race. I said, "You'll have to understand, Freeman Heraal, that it's not our policy to maintain more than a few members of each species at our Institute. And we're not currently in need of any Kallerian males, because—" "You will hire me or trouble I will make!" I opened our inventory chart. I showed him that we were already carrying four Kallerians, and that was more than plenty. The beady little eyes flashed like beacons in the fur. "Yes, you have four representatives—of the Clan Verdrokh! None of the Clan Gursdrinn! For three years, I have waited for a chance to avenge this insult to the noble Clan Gursdrinn!" At the key-word avenge , I readied myself to ensnarl the Kallerian in a spume of tanglemesh the instant he went for his blaster, but he didn't move. He bellowed, "I have vowed a vow, Earthman. Take me to Earth, enroll a Gursdrinn, or the consequences will be terrible!" I'm a man of principles, like all straightforward double-dealers, and one of the most important of those principles is that I never let myself be bullied by anyone. "I deeply regret having unintentionally insulted your clan, Freeman Heraal. Will you accept my apologies?" He glared at me in silence. I went on, "Please be assured that I'll undo the insult at the earliest possible opportunity. It's not feasible for us to hire another Kallerian now, but I'll give preference to the Clan Gursdrinn as soon as a vacancy—" "No. You will hire me now." "It can't be done, Freeman Heraal. We have a budget, and we stick to it." "You will rue! I will take drastic measures!" "Threats will get you nowhere, Freeman Heraal. I give you my word I'll get in touch with you as soon as our organization has room for another Kallerian. And now, please, there are many applicants waiting—" You'd think it would be sort of humiliating to become a specimen in a zoo, but most of these races take it as an honor. And there's always the chance that, by picking a given member of a race, we're insulting all the others. I nudged the trouble-button on the side of my desk and Auchinleck and Ludlow appeared simultaneously from the two doors at right and left. They surrounded the towering Kallerian and sweet-talkingly led him away. He wasn't minded to quarrel physically, or he could have knocked them both into the next city with a backhand swipe of his shaggy paw, but he kept up a growling flow of invective and threats until he was out in the hall. I mopped sweat from my forehead and began to buzz Stebbins for the next applicant. But before my finger touched the button, the door popped open and a small being came scooting in, followed by an angry Stebbins. "Come here, you!" "Stebbins?" I said gently. "I'm sorry, Mr. Corrigan. I lost sight of this one for a moment, and he came running in—" "Please, please," squeaked the little alien pitifully. "I must see you, honored sir!" "It isn't his turn in line," Stebbins protested. "There are at least fifty ahead of him." "All right," I said tiredly. "As long as he's in here already, I might as well see him. Be more careful next time, Stebbins." Stebbins nodded dolefully and backed out. The alien was a pathetic sight: a Stortulian, a squirrely-looking creature about three feet high. His fur, which should have been a lustrous black, was a dull gray, and his eyes were wet and sad. His tail drooped. His voice was little more than a faint whimper, even at full volume. "Begging your most honored pardon most humbly, important sir. I am a being of Stortul XII, having sold my last few possessions to travel to Ghryne for the miserable purpose of obtaining an interview with yourself." I said, "I'd better tell you right at the outset that we're already carrying our full complement of Stortulians. We have both a male and a female now and—" "This is known to me. The female—is her name perchance Tiress?" I glanced down at the inventory chart until I found the Stortulian entry. "Yes, that's her name." The little being immediately emitted a soul-shaking gasp. "It is she! It is she!" "I'm afraid we don't have room for any more—" "You are not in full understanding of my plight. The female Tiress, she is—was—my own Fire-sent spouse, my comfort and my warmth, my life and my love." "Funny," I said. "When we signed her three years ago, she said she was single. It's right here on the chart." "She lied! She left my burrow because she longed to see the splendors of Earth. And I am alone, bound by our sacred customs never to remarry, languishing in sadness and pining for her return. You must take me to Earth!" "But—" "I must see her—her and this disgrace-bringing lover of hers. I must reason with her. Earthman, can't you see I must appeal to her inner flame? I must bring her back! " My face was expressionless. "You don't really intend to join our organization at all—you just want free passage to Earth?" "Yes, yes!" wailed the Stortulian. "Find some other member of my race, if you must! Let me have my wife again, Earthman! Is your heart a dead lump of stone?" It isn't, but another of my principles is to refuse to be swayed by sentiment. I felt sorry for this being's domestic troubles, but I wasn't going to break up a good act just to make an alien squirrel happy—not to mention footing the transportation. I said, "I don't see how we can manage it. The laws are very strict on the subject of bringing alien life to Earth. It has to be for scientific purposes only. And if I know in advance that your purpose in coming isn't scientific, I can't in all conscience lie for you, can I?" "Well—" "Of course not." I took advantage of his pathetic upset to steam right along. "Now if you had come in here and simply asked me to sign you up, I might conceivably have done it. But no—you had to go unburden your heart to me." "I thought the truth would move you." "It did. But in effect you're now asking me to conspire in a fraudulent criminal act. Friend, I can't do it. My reputation means too much to me," I said piously. "Then you will refuse me?" "My heart melts to nothingness for you. But I can't take you to Earth." "Perhaps you will send my wife to me here?" There's a clause in every contract that allows me to jettison an unwanted specimen. All I have to do is declare it no longer of scientific interest, and the World Government will deport the undesirable alien back to its home world. But I wouldn't pull a low trick like that on our female Stortulian. I said, "I'll ask her about coming home. But I won't ship her back against her will. And maybe she's happier where she is." The Stortulian seemed to shrivel. His eyelids closed half-way to mask his tears. He turned and shambled slowly to the door, walking like a living dishrag. In a bleak voice, he said, "There is no hope then. All is lost. I will never see my soulmate again. Good day, Earthman." He spoke in a drab monotone that almost, but not quite, had me weeping. I watched him shuffle out. I do have some conscience, and I had the uneasy feeling I had just been talking to a being who was about to commit suicide on my account. About fifty more applicants were processed without a hitch. Then life started to get complicated again. Nine of the fifty were okay. The rest were unacceptable for one reason or another, and they took the bad news quietly enough. The haul for the day so far was close to two dozen new life-forms under contract. I had just about begun to forget about the incidents of the Kallerian's outraged pride and the Stortulian's flighty wife when the door opened and the Earthman who called himself Ildwar Gorb of Wazzenazz XIII stepped in. "How did you get in here?" I demanded. "Your man happened to be looking the wrong way," he said cheerily. "Change your mind about me yet?" "Get out before I have you thrown out." Gorb shrugged. "I figured you hadn't changed your mind, so I've changed my pitch a bit. If you won't believe I'm from Wazzenazz XIII, suppose I tell you that I am Earthborn, and that I'm looking for a job on your staff." "I don't care what your story is! Get out or—" "—you'll have me thrown out. Okay, okay. Just give me half a second. Corrigan, you're no fool, and neither am I—but that fellow of yours outside is . He doesn't know how to handle alien beings. How many times today has a life-form come in here unexpectedly?" I scowled at him. "Too damn many." "You see? He's incompetent. Suppose you fire him, take me on instead. I've been living in the outworlds half my life; I know all there is to know about alien life-forms. You can use me, Corrigan." I took a deep breath and glanced all around the paneled ceiling of the office before I spoke. "Listen, Gorb, or whatever your name is, I've had a hard day. There's been a Kallerian in here who just about threatened murder, and there's been a Stortulian in here who's about to commit suicide because of me. I have a conscience and it's troubling me. But get this: I just want to finish off my recruiting, pack up and go home to Earth. I don't want you hanging around here bothering me. I'm not looking to hire new staff members, and if you switch back to claiming you're an unknown life-form from Wazzenazz XIII, the answer is that I'm not looking for any of those either. Now will you scram or—" The office door crashed open at that point and Heraal, the Kallerian, came thundering in. He was dressed from head to toe in glittering metalfoil, and instead of his ceremonial blaster, he was wielding a sword the length of a human being. Stebbins and Auchinleck came dragging helplessly along in his wake, hanging desperately to his belt. "Sorry, Chief," Stebbins gasped. "I tried to keep him out, but—" Heraal, who had planted himself in front of my desk, drowned him out with a roar. "Earthman, you have mortally insulted the Clan Gursdrinn!" Sitting with my hands poised near the meshgun trigger, I was ready to let him have it at the first sight of actual violence. Heraal boomed, "You are responsible for what is to happen now. I have notified the authorities and you prosecuted will be for causing the death of a life-form! Suffer, Earthborn ape! Suffer!" "Watch it, Chief," Stebbins yelled. "He's going to—" An instant before my numb fingers could tighten on the meshgun trigger, Heraal swung that huge sword through the air and plunged it savagely through his body. He toppled forward onto the carpet with the sword projecting a couple of feet out of his back. A few driblets of bluish-purple blood spread from beneath him. Before I could react to the big life-form's hara-kiri, the office door flew open again and three sleek reptilian beings entered, garbed in the green sashes of the local police force. Their golden eyes goggled down at the figure on the floor, then came to rest on me. "You are J. F. Corrigan?" the leader asked. "Y-yes." "We have received word of a complaint against you. Said complaint being—" "—that your unethical actions have directly contributed to the untimely death of an intelligent life-form," filled in the second of the Ghrynian policemen. "The evidence lies before us," intoned the leader, "in the cadaver of the unfortunate Kallerian who filed the complaint with us several minutes ago." "And therefore," said the third lizard, "it is our duty to arrest you for this crime and declare you subject to a fine of no less than $100,000 Galactic or two years in prison." "Hold on!" I stormed. "You mean that any being from anywhere in the Universe can come in here and gut himself on my carpet, and I'm responsible?" "This is the law. Do you deny that your stubborn refusal to yield to this late life-form's request lies at the root of his sad demise?" "Well, no, but—" "Failure to deny is admission of guilt. You are guilty, Earthman." Closing my eyes wearily, I tried to wish the whole babbling lot of them away. If I had to, I could pony up the hundred-grand fine, but it was going to put an awful dent in this year's take. And I shuddered when I remembered that any minute that scrawny little Stortulian was likely to come bursting in here to kill himself too. Was it a fine of $100,000 per suicide? At that rate, I could be out of business by nightfall. I was spared further such morbid thoughts by yet another unannounced arrival. The small figure of the Stortulian trudged through the open doorway and stationed itself limply near the threshold. The three Ghrynian policemen and my three assistants forgot the dead Kallerian for a moment and turned to eye the newcomer. I had visions of unending troubles with the law here on Ghryne. I resolved never to come here on a recruiting trip again—or, if I did come, to figure out some more effective way of screening myself against crackpots. In heart-rending tones, the Stortulian declared, "Life is no longer worth living. My last hope is gone. There is only one thing left for me to do." I was quivering at the thought of another hundred thousand smackers going down the drain. "Stop him, somebody! He's going to kill himself! He's—" Then somebody sprinted toward me, hit me amidships, and knocked me flying out from behind my desk before I had a chance to fire the meshgun. My head walloped the floor, and for five or six seconds, I guess I wasn't fully aware of what was going on. Gradually the scene took shape around me. There was a monstrous hole in the wall behind my desk; a smoking blaster lay on the floor, and I saw the three Ghrynian policemen sitting on the raving Stortulian. The man who called himself Ildwar Gorb was getting to his feet and dusting himself off. He helped me up. "Sorry to have had to tackle you, Corrigan. But that Stortulian wasn't here to commit suicide, you see. He was out to get you." I weaved dizzily toward my desk and dropped into my chair. A flying fragment of wall had deflated my pneumatic cushion. The smell of ashed plaster was everywhere. The police were effectively cocooning the struggling little alien in an unbreakable tanglemesh. "Evidently you don't know as much as you think you do about Stortulian psychology, Corrigan," Gorb said lightly. "Suicide is completely abhorrent to them. When they're troubled, they kill the person who caused their trouble. In this case, you." I began to chuckle—more of a tension-relieving snicker than a full-bodied laugh. "Funny," I said. "What is?" asked the self-styled Wazzenazzian. "These aliens. Big blustery Heraal came in with murder in his eye and killed himself , and the pint-sized Stortulian who looked so meek and pathetic damn near blew my head off." I shuddered. "Thanks for the tackle job." "Don't mention it," Gorb said. I glared at the Ghrynian police. "Well? What are you waiting for? Take that murderous little beast out of here! Or isn't murder against the local laws?" "The Stortulian will be duly punished," replied the leader of the Ghrynian cops calmly. "But there is the matter of the dead Kallerian and the fine of—" "—one hundred thousand dollars. I know." I groaned and turned to Stebbins. "Get the Terran Consulate on the phone, Stebbins. Have them send down a legal adviser. Find out if there's any way we can get out of this mess with our skins intact." "Right, Chief." Stebbins moved toward the visiphone. Gorb stepped forward and put a hand on his chest. "Hold it," the Wazzenazzian said crisply. "The Consulate can't help you. I can." "You?" I said. "I can get you out of this cheap." " How cheap?" Gorb grinned rakishly. "Five thousand in cash plus a contract as a specimen with your outfit. In advance, of course. That's a heck of a lot better than forking over a hundred grand, isn't it?" I eyed Gorb uncertainly. The Terran Consulate people probably wouldn't be much help; they tried to keep out of local squabbles unless they were really serious, and I knew from past experiences that no officials ever worried much about the state of my pocketbook. On the other hand, giving this slyster a contract might be a risky proposition. "Tell you what," I said finally. "You've got yourself a deal—but on a contingency basis. Get me out of this and you'll have five grand and the contract. Otherwise, nothing." Gorb shrugged. "What have I to lose?" | B. He was able to offer him less, knowing he would still accept and be grateful. |
How does the method measure the impact of the event on market prices? | ### Introduction
The financial performance of a corporation is correlated with its social responsibility such as whether their products are environmentally friendly, manufacturing safety procedures protect against accidents, or they use child labors in its third world country factories. Consumers care about these factors when making purchasing decisions in the supermarkets and investors integrate environmental, social and governance factors, known as ESG, in their investment decision-making. It has been shown that corporations financial results have a positive correlation with their sustainability business model and the ESG investment methodology can help reduce portfolio risk and generate competitive returns. However, one barrier for ESG evaluation is the lack of relatively complete and centralized information source. Currently, ESG analysts leverage financial reports to collect the the necessary data for proper evaluation such as greenhouse gas emissions or discrimination lawsuits, but this data is inconsistent and latent. In this study, we consider social media a crowdsourcing data feed to be a new data source for this task. Social media applications such as Twitter offer users a platform to share and disseminate almost any content about various events such as sports, music, and controversial events as well. The content produced through these platforms not only facilitates the spread of information but can also provides meaningful signals about the influence of the events. A large number of responses to an issue on Twitter could inform the public about the significance of an event, widen the scope of the event, and bring more public attention inside and outside the social media circle. We define a controversial event for a business entity as a credible and newsworthy incident that has the potential to impact an entity in its financial performance and operation, for example, an incident caused by an employee or a representative of the entity that has the potential to hurt the trust of the public to its brand. Such an incident can demonstrate a potential gap in its risk management framework and policy execution, and eventually hurt the interest and trust of its stakeholders'. Controversial events trigger a large cascade of discussion on social media platforms. The broad connectivity between people propagates their opinions into trending topics that could effect the company financially and operationally. In certain cases, the responsible entity can be forced to take actions, e.g., to recall its product, which can impose a large financial burden on the entity. For instance, in the Takata air bag scandal, the event was discussed widely on Twitter after the New York Times published a comprehensive article on its defective air bag products in 2014. Takata was forced to recall nearly 50 million air bag and filed bankruptcy in June 2017. To this end, we propose a controversial event detection system utilizing Twitter data. We focus on controversial events which are credible and newsworthy. Twitter data were collected on a given company and various attributes of each tweet were extracted. We verify the credibility of the event by validating the URLs appearing in tweets come from credible news sources. We utilize tweets attributes to detect events specific to the given company and the sentiment of the event to measure the controversy. Relationship between a burst of an entity controversial event and the entity market performance data was qualitatively assessed in our case study, where we found its potential impact on the equity value. ### Related Work
There have been a few studies on assessing sustainability of entities. The UN Commission on Sustainable Development (CSD) published a list of about 140 indicators on various dimensions of sustainability BIBREF0 . In BIBREF1 , Singh et al. reviewed various methodologies, indicators, and indices on sustainability assessment, which includes environmental and social domains. All the data, on which the assessments were conducted, mentioned in their works are processed datasets, and some of them are collected from company annual reports and publications, newspaper clips, and management interviews. They stated that the large number of indicators or indices raises the need of data collection. Our work uses the social media data as a new alternative data source to complement the traditional data collection. Event detection on social media has been a popular research topic for years. Reuters Tracer BIBREF2 is reported as an application built for the journalists to detect news leads in Twitter before the news becomes known to the public. Petrovic et al. BIBREF3 presented a locality-sensitive hashing based first story detection algorithm with a new variance reduction strategy to improve the performance. In BIBREF4 , the signal of a tweet word is built with wavelet analysis and a event is detected by clustering words with similar signal patterns of burst. BIBREF5 describes a detection and analysis system named TEDAS which concentrates on Crime and Disaster related Events (CDE). TEDAS classifies if a tweet is a CDE tweet, predicts its geo-location if missing, and ranks and returns important tweets when user queries in the system. TEDAS treats a tweet as an event if the tweet qualifies, while our definition of an event is different, where an event is a group of tweets discussing a same theme. ### Controversy Detection in Social Media
In this section, we describe the main components of our controversy detection system. ### Data collection
The system uses Twitter's filtered streaming API to collect relevant tweets data. The data collection pipeline accepts a comma-separated list of phrases as filtering parameters, that the API uses to determine which tweets will be retained from the stream. Once the system receives data from the API, it then separates postings by companies and runs the downstream process on the separated data streams individually. ### Feature engineering
The data collection pipeline collects tweet postings for a given entity. For each incoming posting, the system also stores the following attributes: posting_id, creation_time, text, language, source, URLs, and hashtags. The system parses the text attribute of each tweet. Part-of-speech (POS) tagging and named entity recognition (NER) algorithm are applied to each tweet and terms that are tagged as proper nouns, verbs, and entities are stored. If two proper nouns are next to each other, the system merges them as one proper noun phrase. Entities such as person names, organizations, locations from tweets are the key elements in describing an event and distinguishing it from other events, and are often used by news professionals to describe the complete story of an event. The verbs from POS tagging mainly represent what and why information, while NER helps to identify where, when, and who information. They capture the major aspects of an event, named who, what, where, when, and why (5W). Besides that, the sentiment of each tweet is assessed too. The system crawls the URLs in a posting and verifies whether the link comes from one or more credible news sources. More specifically, the system may consider the following to be examples of credible news sources: 1) a news outlet that has, and consistently applies, journalistic standards in its reporting or 2) an authoritative government agency not acting in a political capacity. Determining whether a source is a credible news source depends on the context of the event. Based on all the extracted features, the system can build a tweet vector, which includes the following features: tweet id, creation time, source, hashtags, entity/proper nouns, verbs, sentiment, and news links. ### Event detection
When a new tweet is received in the data pipeline, it either forms a new cluster or it will be added to an existing cluster. A new tweet will be added to an existing cluster if it is sufficiently similar to one of the existing clusters based on its distance to the cluster average vector. If more than one cluster is applicable, the cluster that has the highest similarity to the new tweet is picked. If a new tweet is not added to any existing clusters, it would form a new cluster. A candidate event is a cluster that has at least five tweets. Algorithm SECREF6 summarizes our event detection method and the following controversy identification method. ### Controversy identification
An event can be controversial if the public expresses dissenting opinions, usually associated with negative sentiments to it. The system filters out irrelevant events and noise from the established controversial events using the following metrics: The burstiness of an event: To detect the burstiness of an event, the system detects the volume of tweets per time period, e.g., per day, for the entity in question. An event is flagged when the velocity of the volume increase exceeds a threshold. Newsworthiness detection: The system counts the total number of unique verified news links in each cluster and log that count as a newsworthiness metric. Sentiment: For each cluster, its overall sentiment score is quantified by the mean of the sentiment scores among all tweets. Candidate events are ranked based on these metrics, and high ranked events are considered controversial events. Outline of the controversy detection algorithm [1] INLINEFORM0 is a stream of tweets about company INLINEFORM1 Controversy INLINEFORM0 INLINEFORM0 (event detection) INLINEFORM1 TweetFeature INLINEFORM2 INLINEFORM0 INLINEFORM1 INLINEFORM2 current event clusters INLINEFORM3 ClusterFeature INLINEFORM4 INLINEFORM5 Distance INLINEFORM6 compute distance INLINEFORM7 INLINEFORM0 find the closet cluster INLINEFORM1 INLINEFORM2 INLINEFORM3 is merge threshold merge INLINEFORM4 in INLINEFORM5 INLINEFORM6 INLINEFORM7 is singleton cluster INLINEFORM0 INLINEFORM1 is min cluster size as event INLINEFORM0 (controversy identification) INLINEFORM1 Bustiness INLINEFORM2 INLINEFORM3 Newsworthiness INLINEFORM4 INLINEFORM5 INLINEFORM6 SentimentClassify INLINEFORM7 INLINEFORM0 AVG INLINEFORM1 compute event level sentiment INLINEFORM2 combined controversy score INLINEFORM0 INLINEFORM1 is controversial events set ### Case Study - Starbucks Controversy
In this section, we provide a case study of our model on a Starbucks controversial event captured in the system. We validated the event with the Wikipedia page of Starbucks and the major new agencies reports. After the event was detected, its impact was further assessed by linking to the market equity data. On April 12th, 2018, an incident occurred in a Starbucks in Philadelphia, PA. Two African-American men were arrested by the police officials inside that Starbucks. It was reported that the two were denied to access the restroom by the store staff because they did not make any purchase. While waiting at the table, they were told by the staff to leave as they were not making any purchase. They did not comply and thus the store manager called the police and reported that they are trespassing. The two were arrested by the officials but released afterwards without any pressed charges. The scene of the arresting was posted on Twitter and quickly garnered public attention. The video had been viewed more than three millions times in a couple of days and the major local and national news agencies like CNN, NPR, and NYTIMES followed the development of the story. The public outrage originating from the social media universe swiftly triggered a series of chain reaction in the physical world. Protesters gathered together inside and outside the Starbucks store to demand the manager be fired. Several days later, the CEO of the Starbucks issued a public apology for the incident on an ABC's program and stated that he would like to meet the men to show them compassion. To remedy the bad outcome of the event, Starbucks closed its 8,000 stores in the U.S. on May 29th for racial-bias training for its 175K employees. A financial settlement was also established between the two men and Starbucks corporation. This event garnered a serious public relations crisis for Starbucks. Figure FIGREF10 shows the event clusters for six days sampled between April 10th and April 20th. Given the difficulty in showing all of the tweets that were clustered, we use the volume of key POS tagged words (5Ws) detected in the cluster of tweets to approximate the event content. The keywords on the top of each bar reveal aspects of the event cluster. This controversial Starbucks event was captured in our system on April 13th, one day after the event occurred. Prior to the event, the discussion themes about Starbucks (clusters) on Twitter were more random and included topics such as Starbucks gift card, barista, coffee as shown on 04/11/2018. The size of the clusters and the total volume of the tweets per day is comparably small. The first event cluster the system detected associates with the keyword `black', where twitter users mentioned `[...] arrested for being Black'. After the event, the volume of the tweets per day surged multiple times more than before and kept climbing for about a week as the event was developing. The system clearly uncovers the events by being able to pinpoint the clustering keywords `black men', `philly', `CEO', `close', etc. The sentiment scores of the discussion in the clusters for each day are shown on the top part of Figure FIGREF10 . The sentiment score is in a range of -2 to +2, -2 standing for very negative, 0 for neutral, and +2 for very positive. As the figure shows, twitter users' attitude turned from neutral to negative post the Starbucks event occurrence. The quick turn of sentiment polarity serves as an measurement of the event being controversy. Through the validation of the domain of the URLs quoted in the clustered tweets, the authentication of the event is verified. All of the elements of this event indicate that a controversy, specifically, a social related controversy, has occurred. We also did a qualitative study on the Starbucks (SBUX) stock movement during this event. Figure FIGREF12 is the daily percentage change of SBUX and NASDAQ index between April 11th and April 20th. SBUX did not follow the upward trend of the whole market before April 17th, and then its change on April 20th, INLINEFORM0 , is quite significant from historical norms. We collected the historical 52 week stock prices prior to this event and calculated the daily stock price change. The distribution of the daily price change of the previous 52 weeks is Figure FIGREF13 with a mean INLINEFORM1 and standard deviation INLINEFORM2 . The INLINEFORM3 down almost equals to two standard deviations below the mean. Our observation is that plausibly, there was a negative aftereffect from the event of the notable decline in Starbucks stock price due to the major public relations crisis. ### Conclusions
We present the use of Twitter as a new data source to detect controversial events for business entities. Each tweet is represented by a vector comprising name entities and verbs mentioned in the raw tweet text. Events can be identified by grouping similar tweets in the vector space, the size and burstiness of the event, and the sentiment polarities. This system is a data-driven controversy monitoring tool that sifts through large volumes of Twitter data. It provides investors with data on key insights on social consciousness, which allows investors to make more informed investment decisions. The direction of our future work is to: 1) develop a quantitative measure on the event impact on the equity market; 2) identify the relevance of the events to entities' operations; 3) extract post-event mitigation actions from the entities. Figure 1: Event clusters and the sentiment polarity score along the timeline. Figure 3: Histogram of Starbucks stock price daily changes Figure 2: Starbucks stock price and NASDAQ index between April 11th 2018 and April 20th 2018. | We collected the historical 52 week stock prices prior to this event and calculated the daily stock price change. The distribution of the daily price change of the previous 52 weeks is Figure FIGREF13 with a mean INLINEFORM1 and standard deviation INLINEFORM2 . |
Regarding Mr. Chapman, what was the finding in the CT Head on 04/21/2017?
Choose the correct answer from the following options:
A. Ischemic stroke
B. Brain tumor
C. Malresorptive hydrocephalus
D. Cerebral abscess
E. Normal study
| ### Patient Report 0
**Dear colleague, **
We are reporting on our shared patient, Mr. John Chapman, born on
11/16/1994, who received emergency treatment at our clinic on
04/03/2017.
**Diagnoses**:
- Severe open traumatic brain injury with fractures of the cranial
vault, mastoid, and skull base
- Dissection of the distal internal carotid artery on both sides
- Subarachnoid hemorrhage involving both hemispheres and extending
into the basal cisterns
- Aspiration pneumonia
**Other Diagnoses: **
- Status post rib fracture 2005
- Status post appendectomy 2006
- Status post distal radius fracture 2008
- Status post elbow fracture 20010
**Procedure**: External ventricular drain (EVD) placement.
**Medical History:** Admission through the emergency department as a
polytrauma alert. The patient was involved in a motocross accident,
where he jumped, fell, and landed face-first. He was intubated at the
scene, and either during or before intubation, aspiration occurred. No
issues with airway, breathing, or circulation (A, B, or C problems) were
noted. A CT scan performed in the emergency department revealed an open
traumatic brain injury with fractures of the cranial vault, mastoid, and
skull base, as well as dissection of both carotid arteries. Upon
admission, we encountered an intubated and sedated patient with a
Richmond Agitation-Sedation Scale (RASS) score of -4. He was
hemodynamically stable at all times.
**Current Recommendations:**
- Regular checks of vigilance, laboratory values and microbiological
findings.
- Careful balancing
### Patient Report 1
**Dear colleague, **
We report on Mr. John Chapman, born on 11/16/1994, who was admitted to
our Intensive Care Unit from 04/03/2017 to 05/01/2017.
**Diagnoses:**
- Open severe traumatic brain injury with fractures of the skull
vault, mastoid, and skull base
- Dissection of the distal ACI on both sides
- Subarachnoid hemorrhage involving both hemispheres and extending
into basal cisterns
- Infarct areas in the border zone between MCA-ACA on the right
frontal and left parietal sides
- Malresorptive hydrocephalus
<!-- -->
- Rhabdomyolysis
- Aspiration pneumonia
**Other Diagnoses: **
- Status post rib fracture in 2005
- Status post appendectomy in 2006
- Status post distal radius fracture in 2008
- Status post elbow fracture in 20010
**Surgical Procedures:**
- 04/03/2017: Placement of external ventricular drain
- 04/08/2017: Placement of an intracranial pressure monitoring
catheter
- 04/13/2017: Surgical tracheostomy
- 05/01/2017: Left ventriculoperitoneal shunt placement
**Medical History:** The patient was admitted through the emergency
department as a polytrauma alert. The patient had fallen while riding a
motocross bike, landing face-first after jumping. He was intubated at
the scene. Aspiration occurred either during or before intubation. No
problems with breathing or circulation were noted. The CT performed in
the emergency department showed an open traumatic brain injury with
fractures of the skull vault, mastoid, and skull base, as well as
dissection of the carotid arteries on both sides and bilateral
subarachnoid hemorrhage.
Upon admission, the patient was sedated and intubated, with a Richmond
Agitation-Sedation Scale (RASS) score of -4, and was hemodynamically
stable under controlled ventilation.
**Therapy and Progression:**
[Neurology]{.underline}: Following the patient\'s admission, an external
ventricular drain was placed. Reduction of sedation had to be
discontinued due to increased intracranial pressure. A right pupil size
greater than the left showed no intracranial correlate. With
persistently elevated intracranial pressure, intensive intracranial
pressure therapy was initiated using deeper sedation, administration of
hyperosmolar sodium, and cerebrospinal fluid drainage, which normalized
intracranial pressure. Intermittently, there were recurrent intracranial
pressure peaks, which could be treated conservatively. Transcranial
Doppler examinations showed normal flow velocities. Microbiological
samples from cerebrospinal fluid were obtained when the patient had
elevated temperatures, but no bacterial growth was observed. Due to the
inability to adequately monitor intracranial pressure via the external
ventricular drain, an intracranial pressure monitoring catheter was
placed to facilitate adequate intracranial pressure monitoring. In the
perfusion computed tomography, progressive edema with increasingly
obstructed external ventricular spaces and previously known infarcts in
the border zone area were observed. To ensure appropriate intracranial
pressure monitoring, a Tuohy drain was inserted due to cerebrospinal
fluid buildup on 04/21/2017. After the initiation of antibiotic therapy
for suspected ventriculitis, the intracranial pressure monitoring
catheter was removed on 04/20/2017. Subsequently, a liquorrhea
developed, leading to the placement of a Tuohy drain. After successful
antibiotic treatment of ventriculitis, a ventriculoperitoneal shunt was
placed on 05/01/2017 without complications, and the Tuohy drain was
removed. Radiological control confirmed the correct positioning. The
patient gradually became more alert. Both pupils were isochoric and
reacted to light. All extremities showed movement, although the patient
only intermittently responded to commands. On 05/01/2017, a VP shunt was
placed on the left side without complications. Currently, the patient is
sedated with continuous clonidine at 60µg/h.
**Hemodynamics**: To maintain cerebral perfusion pressure in the
presence of increased intracranial pressure, circulatory support with
vasopressors was necessary. Echocardiography revealed preserved cardiac
function without wall motion abnormalities or right heart strain,
despite the increasing need for noradrenaline support. As the patient
had bilateral carotid dissection, a therapy with Aspirin 100mg was
initiated. On 04/16/2017, clinical examination revealed right\>left leg
circumference difference and redness of the right leg. Utrasound
revealed a long-segment deep vein thrombosis in the right leg, extending
from the pelvis (proximal end of the thrombus not clearly delineated) to
the lower leg. Therefore, Heparin was increased to a therapeutic dose.
Heparin therapy was paused on postoperative day 1, and prophylactic
anticoagulation started, followed by therapeutic anticoagulation on
postoperative day 2. The patient was switched to subcutaneous Lovenox.
**Pulmonary**: Due to the history of aspiration in the prehospital
setting, a bronchoscopy was performed, revealing a moderately obstructed
bronchial system with several clots. As prolonged sedation was
necessary, a surgical tracheostomy was performed without complications
on 04/13/2017. Subsequently, we initiated weaning from mechanical
ventilation. The current weaning strategy includes 12 hours of
synchronized intermittent mandatory ventilation (SIMV) during the night,
with nighttime pressure support ventilation (DuoPAP: Ti high 1.3s,
respiratory rate 11/min, Phigh 11 mbar, PEEP 5 mbar, Psupport 5 mbar,
trigger 4l, ramp 50 ms, expiratory trigger sensitivity 25%).
**Abdomen**: FAST examinations did not reveal any signs of
intra-abdominal trauma. Enteral feeding was initiated via a gastric
tube, along with supportive parenteral nutrition. With forced bowel
movement measures, the patient had regular bowel movements. On
04/17/2017, a complication-free PEG (percutaneous endoscopic
gastrostomy) placement was performed due to the potential long-term need
for enteral nutrition. The PEG tube is currently being fed with tube
feed nutrition, with no bowel movement for the past four days.
Additionally, supportive parenteral nutrition is being provided.
**Kidney**: Initially, the patient had polyuria without confirming
diabetes insipidus, and subsequently, adequate diuresis developed.
Retention parameters were within the normal range. As crush parameters
increased, a therapy involving forced diuresis was initiated, resulting
in a significant reduction of crush parameters.
**Infection Course:** Upon admission, with elevated infection parameters
and intermittently febrile temperatures, empirical antibiotic therapy
was initiated for suspected pneumonia using Piperacillin/Tazobactam.
Staphylococcus capitis was identified in blood cultures, and
Staphylococcus aureus was found in bronchial lavage. Both microbes were
sensitive to the current antibiotic therapy, so treatment with
Piperacillin/Tazobactam continued. Additionally, Enterobacter cloacae
was identified in tracheobronchial secretions during the course, also
sensitive to the ongoing antibiotic therapy. On 05/17, the patient
experienced another fever episode with elevated infection parameters and
right lower lobe infiltrates in the chest X-ray. After obtaining
microbiological samples, antibiotic therapy was switched to Meropenem
for suspected pneumonia. Microbiological findings from cerebrospinal
fluid indicated gram-negative rods. Therefore, antibiotic therapy was
adjusted to Ciprofloxacin in accordance with susceptibility testing due
to suspected ventriculitis, and the Meropenem dose was increased. This
led to a reduction in infection parameters. Finally, microbiological
examination of cerebrospinal fluid, blood cultures, and urine revealed
no pathological findings. Infection parameters decreased. We recommend
continuing antibiotic therapy until 05/02/2017.
**Anti-Infective Course: **
- Piperacillin/Tazobactam 04/03/2017-04/16/2017: Staph. Capitis in
Blood Culture Staph. Aureus in Bronchial Lavage
- Meropenem 04/16/2017-present (increased dose since 04/18) CSF:
gram-negative rods in Blood Culture: Pseudomonas aeruginosa
Acinetobacter radioresistens
- Ciprofloxacin 04/18/2017-present CSF: gram-negative rods in Blood
Culture: Pseudomonas aeruginosa, Acinetobacter radioresistens
**Weaning Settings:** Weaning Stage 6: 12-hour synchronized intermittent
mandatory ventilation (SIMV) with DuoPAP during the night (Thigh 1.3s,
respiratory rate 11/min, Phigh 11 mbar, PEEP 5 mbar, Psupport 5 mbar,
trigger 4l, ramp 50 ms, expiratory trigger sensitivity 25%).
**Status at transfer:** Currently, Mr. Chapman is monosedated with
Clonidine. He spontaneously opens both eyes and spontaneously moves all
four extremities. Pupils are bilaterally moderately dilated, round and
sensitive to light. There is bulbar divergence. Circulation is stable
without catecholamine therapy. He is in the process of weaning,
currently spontaneous breathing with intermittent CPAP. Renal function
is sufficient, enteral nutrition via PEG with supportive parenteral
nutrition is successful.
**Current Medication:**
**Medication** **Dosage** **Frequency**
------------------------------------ ---------------- ---------------
Bisoprolol (Zebeta) 2.5 mg 1-0-0
Ciprofloxacin (Cipro) 400 mg 1-1-1
Meropenem (Merrem) 4 g Every 4 hours
Morphine Hydrochloride (MS Contin) 10 mg 1-1-1-1-1-1
Polyethylene Glycol 3350 (MiraLAX) 13.1 g 1-1-1
Acetaminophen (Tylenol) 1000 mg 1-1-1-1
Aspirin 100 mg 1-0-0
Enoxaparin (Lovenox) 30 mg (0.3 mL) 0-0-1
Enoxaparin (Lovenox) 70 mg (0.7 mL) 1-0-1
**Lab results:**
**Parameter** **Results** **Reference Range**
-------------------- ------------- ---------------------
Creatinine (Jaffé) 0.42 mg/dL 0.70-1.20 mg/dL
Urea 31 mg/dL 17-48 mg/dL
Total Bilirubin 0.35 mg/dL \< 1.20 mg/dL
Hemoglobin 7.6 g/dL 13.5-17.0 g/dL
Hematocrit 28% 39.5-50.5%
Red Blood Cells 3.5 M/uL 4.3-5.8 M/uL
White Blood Cells 10.35 K/uL 3.90-10.50 K/uL
Platelets 379 K/uL 150-370 K/uL
MCV 77.2 fL 80.0-99.0 fL
MCH 24.1 pg 27.0-33.5 pg
MCHC 32.5 g/dL 31.5-36.0 g/dL
MPV 11.3 fL 7.0-12.0 fL
RDW-CV 17.7% 11.5-15.0%
Quick 54% 78-123%
INR 1.36 0.90-1.25
aPTT 32.8 sec 25.0-38.0 sec
**Addition: Radiological Findings**
[Clinical Information and Justification:]{.underline} Suspected deep
vein thrombosis (DVT) on the right leg.
[Special Notes:]{.underline} Examination at the bedside in the intensive
care unit, no digital image archiving available.
[Findings]{.underline}: Confirmation of a long-segment deep venous
thrombosis in the right leg, starting in the pelvis (proximal end not
clearly delineated) and extending to the lower leg.
Visible Inferior Vena Cava without evidence of thrombosis.
The findings were communicated to the treating physician.
**Full-Body Trauma CT on 04/03/2017:**
[Clinical Information and Justification:]{.underline} Motocross
accident. Polytrauma alert. Consequences of trauma? Informed consent:
Emergency indication. Recommended monitoring of kidney and thyroid
laboratory parameters.
**Findings**: CCT: Dissection of the distal internal carotid artery on
both sides (left 2-fold).
Signs of generalized elevated intracranial pressure.
Open skull-brain trauma with intracranial air inclusions and skull base
fracture at the level of the roof of the ethmoidal/sphenoidal sinuses
and clivus (in a close relationship to the bilateral internal carotid
arteries) and the temporal
**CT Head on 04/16/2017:**
[Clinical Information and Justification:]{.underline} History of skull
fracture, removal of EVD (External Ventricular Drain). Inquiry about the
course.
[Findings]{.underline}: Regression of ventricular system width (distance
of SVVH currently 41 mm, previously 46 mm) with residual liquor caps,
indicative of regressed hydrocephalus. Interhemispheric fissure in the
midline. No herniation.
Complete regression of subdural hematoma on the left, tentorial region.
Known defect areas on the right frontal lobe where previous catheters
were inserted.
Progression of a newly hypodense demarcated cortical infarct on the
left, postcentral.
Known bilateral skull base fractures involving the petrous bone, with
secretion retention in the mastoid air cells bilaterally. Minimal
secretion also in the sphenoid sinuses.
Postoperative bone fragments dislocated intracranially after right
frontal trepanation.
**Chest X-ray on 04/24/2017.**
[Clinical Information and Justification:]{.underline} Mechanically
ventilated patient. Suspected pneumonia. Question about infiltrates.
[Findings]{.underline}: Several previous images for comparison, last one
from 08/20/2021.
Persistence of infiltrates in the right lower lobe. No evidence of new
infiltrates. Removal of the tracheal tube and central venous catheter
with a newly inserted tracheal cannula. No evidence of pleural effusion
or pneumothorax.
**CT Head on 04/25/2017:**
[Clinical Information and Justification:]{.underline} Severe traumatic
brain injury with brain edema, one External Ventricular Drain removed,
one parenchymal catheter removed; Follow-up.
[Findings]{.underline}: Previous images available, CT last performed on
04/09/17, and MRI on 04/16/17.
Massive cerebrospinal fluid (CSF) stasis supra- and infratentorially
with CSF pressure caps at the ventricular and cisternal levels with
completely depleted external CSF spaces, differential diagnosis:
malresorptive hydrocephalus. The EVD and parenchymal catheter have been
completely removed.
No evidence of fresh intracranial hemorrhage. Residual subdural hematoma
on the left, tentorial. Slight regression of the cerebellar tonsils.
Increasing hypodensity of the known defect zone on the right frontal
region, differential diagnosis: CSF diapedesis. Otherwise, the status is
the same as for the other defects.
Secretion in the sphenoid sinus and mastoid cells bilaterally, known
bilateral skull base fractures.
**Bedside Chest X-ray on 04/262017:**
[Clinical Information and Justification]{.underline}: Respiratory
insufficiency. Inquiry about cardiorespiratory status.
[Findings]{.underline}: Previous image from 08/17/2021.
Left Central Venous Catheter and gastric tube in unchanged position.
Persistent consolidation in the right para-hilar region, differential
diagnosis: contusion or partial atelectasis. No evidence of new
pulmonary infiltrates. No pleural effusion. No pneumothorax. No
pulmonary congestion.
**Brain MRI on 04/26/2017:**
[Clinical Information and Justification:]{.underline} Severe skull-brain
trauma with skull calvarium, mastoid, and skull base fractures.
Assessment of infarct areas/edema for rehabilitation planning.
[Findings:]{.underline} Several previous examinations available.
Persistent small sulcal hemorrhages in both hemispheres (left \> right)
and parenchymal hemorrhage on the left frontal with minimal perifocal
edema.
Narrow subdural hematoma on the left occipital extending tentorially (up
to 2 mm).
No current signs of hypoxic brain damage. No evidence of fresh ischemia.
Slightly regressed ventricular size. No herniation. Unchanged placement
of catheters on the right frontal side. Mastoid air cells blocked
bilaterally due to known bilateral skull base fractures, mucosal
swelling in the sphenoid and ethmoid sinuses. Polypous mucosal swelling
in the left maxillary sinus. Other involved paranasal sinuses and
mastoids are clear.
**Bedside Chest X-ray on 04/27/2017:**
[Clinical Information and Justification:]{.underline} Tracheal cannula
placement. Inquiry about the position.
[Findings]{.underline}: Images from 04/03/2017 for comparison.
Tracheal cannula with tip projecting onto the trachea. No pneumothorax.
Regressing infiltrate in the right lower lung field. No leaking pleural
effusions.
Left ubclavian central venous catheter with tip projecting onto the
superior vena cava. Gastric tube in situ.
**CT Head on 04/28/2017:**
[Clinical Information and Justification:]{.underline} Open head injury,
bilateral subarachnoid hemorrhage (SAH), EVD placement. Inquiry about
herniation.
[Findings]{.underline}: Comparison with the last prior examination from
the previous day.
Generalized signs of cerebral edema remain constant, slightly
progressing with a somewhat increasing blurred cortical border,
particularly high frontal.
Essentially constant transtentorial herniation of the midbrain and low
position of the cerebellar tonsils. Marked reduction of inner CSF spaces
and depleted external CSF spaces, unchanged position of the ventricular
drainage catheter with the tip in the left lateral ventricle.
Constant small parenchymal hemorrhage on the left frontal and constant
SDH at the tentorial edge on both sides. No evidence of new intracranial
space-occupying hemorrhage.
Slightly less distinct demarcation of the demarcated infarcts/defect
zones, e.g., on the right frontal region, differential diagnosis:
fogging.
**CT Head Angiography with Perfusion on 04/28/2017:**
[Clinical Information and Justification]{.underline}: Post-traumatic
head injury, rising intracranial pressure, bilateral internal carotid
artery dissection. Inquiry about intracranial bleeding, edema course,
herniation, brain perfusion.
[Emergency indication:]{.underline} Vital indication. Recommended
monitoring of kidney and thyroid laboratory parameters. Consultation
with the attending physician from and the neuroradiology service was
conducted.
[Technique]{.underline}: Native moderately of the neurocranium. CT
angiography of brain-supplying cervical intracranial vessels during
arterial contrast agent phase and perfusion imaging of the neurocranium
after intravenous injection of a total of 140 ml of Xenetix-350. DLP
Head 502.4 mGy*cm. DLP Body 597.4 mGy*cm.
[Findings]{.underline}: Previous images from 08/11/2021 and the last CTA
of the head/neck from 04/03/2017 for comparison.
[Brain]{.underline}: Constant bihemispheric and cerebellar brain edema
with a slit-like appearance of the internal and completely compressed
external ventricular spaces. Constant compression of the midbrain with
transtentorial herniation and a constant tonsillar descent.
Increasing demarcation of infarct areas in the border zone of MCA-ACA on
the right frontal, possibly also on the left frontal. Predominantly
preserved cortex-gray matter contrast, sometimes discontinuous on both
frontal sides, differential diagnosis: artifact-related, differential
diagnosis: disseminated infarct demarcations/contusions.
Unchanged placement of the ventricular drainage from the right frontal
with the catheter tip in the left lateral ventricle anterior horn.
Constant subdural hematoma tentorial and posterior falx. Increasingly
vague delineation of the small frontal parenchymal hemorrhage. No new
space-occupying intracranial bleeding.
No evidence of secondary dislocation of the skull base fracture with
constant fluid collections in the paranasal sinuses and mastoid air
cells. Hematoma possible, cerebrospinal fluid leakage possible.
[CT Angiography Head/Neck]{.underline}: Constant presentation of
bilateral internal carotid artery dissection.
No evidence of higher-grade vessel stenosis or occlusion of the
brain-supplying intracranial arteries.
Moderately dilated venous collateral circuits in the cranial soft
tissues on both sides, right \> left. Moderately dilated ophthalmic
veins on both sides, right \> left.
No evidence of sinus or cerebral venous thrombosis. Slight perfusion
deficits in the area of the described infarct areas and contusions.
No evidence of perfusion mismatches in the perfusion imaging.
Unchanged presentation of the other documented skeletal segments.
Additional Note: Discussion of findings with the responsible medical
colleagues on-site and by telephone, as well as with the neuroradiology
service by telephone, was conducted.
**CT Head on 04/30/2017:**
[Clinical Information and Justification]{.underline}: Open head injury
following a motorcycle accident.. Inquiry about rebleeding, edema, EVD
displacement.
[Findings and Assessment:]{.underline} CT last performed on 04/05/2017
for comparison.
Constant narrow subdural hematoma on both sides, tentorial and posterior
parasagittal. Constant small parenchymal hemorrhage on the left frontal.
No new intracranial bleeding.
Progressively demarcated infarcts on the right frontal and left
parietal.
Slightly progressive compression of the narrow ventricles as an
indication of progressive edema. Completely depleted external CSF spaces
with the ventricular drain catheter in the left lateral ventricle.
Increasing compression of the midbrain due to transtentorial herniation,
progressive tonsillar descent of 6 mm.
Fracture of the skull base and the petrous part of the temporal bone on
both sides without significant displacement. Hematoma in the mastoid and
sphenoid sinuses and the maxillary sinus.
**CT Head on 05/01/2017:**
[Clinical Information and Justification:]{.underline} Open skull-brain
trauma. Inquiry about CSF stasis, bleeding, edema.
[Findings]{.underline}: CT last performed on 04/05/17 for comparison.
Completely regressed subarachnoid hemorrhages on both sides. Minimal SDH
components on the tentorial edges bilaterally (left more than right,
with a 3 mm margin width). No new intracranial bleeding. Continuously
narrow inner ventricular system and narrow basal cisterns. The fourth
ventricle is unfolded. Narrow external CSF spaces and consistently
swollen gyration with global cerebral edema.
Better demarcated circumscribed hypodensity in the centrum semiovale on
the right (Series 3, Image 176) and left (Series 3, Image 203);
Differential diagnosis: fresh infarcts due to distal ACI dissections.
Consider repeat vascular imaging. No midline shift. No herniation.
Regressing intracranial air inclusions. Fracture of the skull base and
the petrous part of the temporal bone on both sides without significant
displacement. Hematoma in the maxillary, sphenoidal, and ethmoidal
sinuses.
**Consultation Reports:**
**1) Consultation with Ophthalmology on 04/03/2017**
[Patient Information:]{.underline}
- Motorbike accident, heavily contaminated eyes.
- Request for assessment.
**Diagnosis:** Motorbike accident
**Findings:** Patient intubated, unresponsive. In cranial CT, the
eyeball appears intact, no retrobulbar hematoma. Intraocular pressure:
Right/left within the normal range. Eyelid margins of both eyes crusty
with sand, inferiorly in the lower lid sac, and on the upper lid with
sand. Lower lid somewhat chemotic. Slight temporal hyperemia in the left
eyelid angle. Both eyes have erosions, small, multiple, superficial.
Lower conjunctival sac clean. Round pupils, anisocoria right larger than
left. Left iris hyperemia, no iris defects in the direct light. Lens
unremarkable. Reduced view of the optic nerve head due to miosis,
somewhat pale, rather sharp-edged, central neuroretinal rim present,
central vessels normal. Left eye, due to narrow pupil, limited view,
optic nerve head not visible, central vessels normal, no retinal
hemorrhages.
**Assessment:** Eyelid and conjunctival foreign bodies removed. Mild
erosions in the lower conjunctival sac. Right optic nerve head somewhat
pale, rather sharp-edged.
**Current Recommendations:**
- Antibiotic eye drops three times a day for both eyes.
- Ensure complete eyelid closure.
**2) Consultation with Craniomaxillofacial (CMF) Surgery on 04/05/2017**
**Patient Information:**
- Motorbike accident with severe open traumatic brain injury with
fractures of the cranial vault, mastoid, and skull base
<!-- -->
- Request for assessment.
- Patient with maxillary fracture.
**Findings:** According to the responsible attending physician,
\"minimal handling in case of decompensating intracranial pressure\" is
indicated. Therefore, currently, a cautious approach is suggested
regarding surgical intervention for the radiologically hardly displaced
maxillary fracture. Re-consultation is possible if there are changes in
the clinical outcome.
**Assessment:** Awaiting developments.
**3) Consultation with Neurology on 04/06/2017**
**Patient Information:**
- Brain edema following a severe open traumatic brain injury with
fractures of the cranial vault, mastoid, and skull base
<!-- -->
- Request for assessment.
- Traumatic subarachnoid hemorrhage, intracranial artery dissection,
and various other injuries.
**Findings:** Patient comatose, intubated, sedated. Isocoric pupils. No
light reaction in either eye. No reaction to pain stimuli for
vestibulo-ocular reflex and oculomotor responses. Babinski reflex
negative.
**Assessment:** Long-term ventilation due to a history of intracerebral
bleeding and skull base fracture. No response to pain stimuli or light
reactions in the eyes.
**Procedure/Therapy Suggestion:** Monitoring of patient condition.
**4) Consultation with ENT on 04/16/2017**
**Patient Information:** Tracheostomy tube change.
**Findings:** Tracheostomy tube change performed. Stoma unremarkable.
Trachea clear up to the bifurcation. Sutures in place.
**Assessment:** Re-consultation on 08/27/2021 for suture removal.
**5) Consultation with Neurology on 04/22/2017**
**Patient Information:** Adduction deficit., Request for assessment.
**Findings:** Long-term ventilation due to a history of intracerebral
bleeding and skull base fracture. Adduction deficit in the right eye and
horizontal nystagmus.
**Assessment:** Suspected mesencephalic lesion due to horizontal
nystagmus, but no diagnostic or therapeutic action required.
**6) Consultation with ENT on 04/23/2017**
**Patient Information:** Suture removal. Request for assessment.
**Findings:** Tracheostomy site unremarkable. Sutures trimmed, and skin
sutures removed.
**Assessment:** Procedure completed successfully.
Please note that some information is clinical and may not include
specific dates or recommendations for further treatment.
**Antibiogram:**
**Antibiotic** **Organism 1 (Pseudomonas aeruginosa)** **Organism 2 (Acinetobacter radioresistens)**
------------------------- ----------------------------------------- -----------------------------------------------
Aztreonam I (4.0) \-
Cefepime I (2.0) \-
Cefotaxime \- \-
Amikacin S (\<=2.0) S (4.0)
Ampicillin \- \-
Piperacillin I (\<=4.0) \-
Piperacillin/Tazobactam I (8.0) \-
Imipenem I (2.0) S (\<=0.25)
Meropenem S (\<=0.25) S (\<=0.25)
Ceftriaxone \- \-
Ceftazidime I (4.0) \-
Gentamicin . (\<=1.0) S (\<=1.0)
Tobramycin S (\<=1.0) S (\<=1.0)
Cotrimoxazole \- S (\<=20.0)
Ciprofloxacin I (\<=0.25) I (0.5)
Moxifloxacin \- \-
Fosfomycin \- \-
Tigecyclin \- \-
\"S\" means Susceptible
\"I\" means Intermediate
\".\" indicates not specified
\"-\" means Resistant
### Patient Report 2
**Dear colleague, **
We are reporting on our mutual patient, Mr. John Chapman, born on
11/16/1994, who presented himself to our Outpatient Clinic from
08/08/2018.
**Diagnoses**:
- Right abducens Nerve Palsy and Facial Nerve Palsy
- Lagophthalmos with corneal opacities due to eyelid closure deficit
- Left Abducens Nerve Palsy with slight compensatory head leftward
rotation and preferred leftward gaze
- Bilateral disc swelling
- Suspected left cavernous internal carotid artery aneurysm following
traumatic ICA dissection
- History of shunt explantation due to dysfunction and right-sided
re-implantation (Codman, current pressure setting 12 cm H2O)
- History of left VP shunt placement (programmable
ventriculoperitoneal shunt, initial pressure setting 5/25 cm H2O,
adjusted to 3 cm H2O before discharge)
- Malresorptive hydrocephalus
- History of severe open head injury in a motocross accident with
multiple skull fractures and distal dissection
**Procedure**: We conducted the following preoperative assessment:
- Visual acuity: Distant vision: Right eye: 0.5, Left eye: 0.8p
- Eye position: Fusion/Normal with significant esotropia in the right
eye; no fusion reflex observed
- Ocular deviation: After CT, at distance, esodeviation simulating
alternating 100 prism diopters (overcorrection); at near,
esodeviation simulating alternating 90 prism diopters
- Head posture: Fusion/Normal with leftward head turn of 5-10 degrees
- Correspondence: Bagolini test shows suppression at both distance and
near fixation
- Motility: Right eye abduction limited to 25 degrees from the
midline, abduction in up and down gaze limited to 30 degrees from
midline; left eye abduction limited to 30 degrees
- Binocular functions: Bagolini test shows suppression in the right
eye at both distance and near fixation; Lang I negative
**Current Presentation:** Mr. Chapman presented himself today in our
neurovascular clinic, providing an MRI of the head.
**Medical History:** The patient is known to have a pseudoaneurysm of
the cavernous left internal carotid artery following traumatic carotid
dissection in 04/2017, along with ipsilateral abducens nerve palsy.
**Physical Examination:** Patient in good general condition. Oriented in
all aspects. No cyanosis. No edema. Warm and dry skin. Normal nasal and
pharyngeal findings. Pupils round, equal, and react promptly to light
bilaterally. Moist tongue. Pharynx and buccal mucosa unremarkable. No
jugular vein distension. No carotid bruits heard. Palpation of lymph
nodes unremarkable. Palpation of the thyroid gland unremarkable, freely
movable. Lungs: Normal chest shape, moderately mobile, vesicular breath
sounds. Heart: Regular heart action, normal rate; heart sounds clear, no
pathological sounds. Abdomen: Peristalsis and bowel sounds normal in all
quadrants; soft abdomen, no tenderness, no palpable masses, liver and
spleen not palpable due to limited access, non-tender kidneys. Normal
peripheral pulses; joints freely movable. Strength, motor function, and
sensation are unremarkable.
**Therapy and Progression:** The pseudoaneurysm has shown slight
enlargement in the recent follow-up imaging and remains partially
thrombosed. The findings were discussed on during a neurovascular board
meeting, where a recommendation for endovascular treatment was made,
which the patient has not yet pursued. Since Mr. Chapman has not been
able to decide on treatment thus far, it is advisable to further
evaluate this still asymptomatic condition through a diagnostic
angiography. This examination would also help in better planning any
potential intervention. Mr. Chapman agreed to this course of action, and
we will provide him with a timely appointment for the angiography.
**Lab results upon Discharge:**
**Parameter** **Results** **Reference Range**
-------------------- ------------- ---------------------
Creatinine (Jaffé) 0.44 mg/dL 0.70-1.20 mg/dL
Urea 31 mg/dL 17-48 mg/dL
Total Bilirubin 0.35 mg/dL \< 1.20 mg/dL
Hemoglobin 7.8 g/dL 13.5-17.0 g/dL
Hematocrit 28% 39.5-50.5%
Red Blood Cells 3.5 M/uL 4.3-5.8 M/uL
White Blood Cells 10.35 K/uL 3.90-10.50 K/uL
Platelets 379 K/uL 150-370 K/uL
MCV 77.2 fL 80.0-99.0 fL
MCH 24.1 pg 27.0-33.5 pg
MCHC 32.5 g/dL 31.5-36.0 g/dL
MPV 11.3 fL 7.0-12.0 fL
RDW-CV 17.7% 11.5-15.0%
Quick 54% 78-123%
INR 1.36 0.90-1.25
aPTT 32.8 sec 25.0-38.0 sec
### Patient Report 3
**Dear colleague, **
We are reporting on our patient, Mr. John Chapman, born on 11/16/1994,
who was under our inpatient care from 05/25/2019 to 05/26/2019.
**Diagnoses: **
- Pseudoaneurysm of the cavernous left internal carotid artery
following traumatic carotid dissection
- Abducens nerve palsy.
- History of severe open head trauma with fractures of the cranial
vault, mastoid, and skull base. Distal ICA dissection bilaterally.
Bilateral hemispheric subarachnoid hemorrhage extending into the
basal cisterns.mInfarct areas in the MCA-ACA border zones, right
frontal, and left parietal. Malresorptive hydrocephalus.
<!-- -->
- Rhabdomyolysis.
- History of aspiration pneumonia.
- Suspected Propofol infusion syndrome.
**Current Presentation:** For cerebral digital subtraction angiography
of the intracranial vessels. The patient presented with stable
cardiopulmonary conditions.
**Medical History**: The patient was admitted for the evaluation of a
pseudoaneurysm of the supra-aortic vessels. Further medical history can
be assumed to be known.
**Physical Examination:** Patient in good general condition. Oriented in
all aspects. No cyanosis. No edema. Warm and dry skin. Normal nasal and
pharyngeal findings. Pupils round, equal, and react promptly to light
bilaterally. Moist tongue. Pharynx and buccal mucosa unremarkable. No
jugular vein distension. No carotid bruits heard. Palpation of lymph
nodes unremarkable. Palpation of the thyroid gland unremarkable, freely
movable. Lungs: Normal chest shape, moderately mobile, vesicular breath
sounds. Heart: Regular heart action, normal rate; heart sounds clear, no
pathological sounds. Abdomen: Peristalsis and bowel sounds normal in all
quadrants; soft abdomen, no tenderness, no palpable masses, liver and
spleen not palpable due to limited access, non-tender kidneys. Normal
peripheral pulses; joints freely movable. Strength, motor function, and
sensation are unremarkable.
**Supra-aortic angiography on 05/25/2019:**
[Clinical context, question, justifying indication:]{.underline}
Pseudoaneurysm of the left ICA. Written consent was obtained for the
procedure. Anesthesia, Medications: Procedure performed under local
anesthesia. Medications: 500 IU Heparin in 500 mL NaCl for flushing.
[Methodology]{.underline}: Puncture of the right common femoral artery
under local anesthesia. 4F sheath, 4F vertebral catheter. Serial
angiographies after selective catheterization of the internal carotid
arteries. Uncomplicated manual intra-arterial contrast medium injection
with a total of 50 mL of Iomeron 300. Post-interventional closure of the
puncture site by manual compression. Subsequent application of a
circular pressure bandage.
[Technique]{.underline}: Biplanar imaging technique, area dose product
1330 cGy x cm², fluoroscopy time 3:43 minutes.
[Findings]{.underline}: The perfused portion of the partially thrombosed
cavernous aneurysm of the left internal carotid artery measures 4 x 2
mm. No evidence of other vascular pathologies in the anterior
circulation.
[Recommendation]{.underline}: In case of post-procedural bleeding,
immediate manual compression of the puncture site and notification of
the on-call neuroradiologist are advised.
- Pressure bandage to be kept until 2:30 PM. Bed rest until 6:30 PM.
- Follow-up in our Neurovascular Clinic
**Addition: Doppler ultrasound of the right groin on 05/26/2019:**
[Clinical context, question, justifying indication:]{.underline} Free
fluid? Hematoma?
[Findings]{.underline}: A CT scan from 04/05/2017 is available for
comparison. No evidence of a significant hematoma or an aneurysm in the
right groin puncture site. No evidence of an arteriovenous fistula.
Normal flow profiles of the femoral artery and vein. No evidence of
thrombosis.
**Treatment and Progression:** Pre-admission occurred on 05/24/2019 due
to a medically justified increase in risk for DSA of intracranial
vessels. After appropriate preparation, the angiography was performed on
05/25/2019. The puncture site was managed with a pressure bandage. In
the color Doppler sonographic control the following day, neither a
puncture aneurysm nor an arteriovenous fistula was detected. On
05/25/2019, we discharged the patient in good subjective condition for
your outpatient follow-up care.
**Current Recommendations:** Outpatient follow-up
**Lab results:**
**Parameter** **Reference Range** **Result**
----------------------- --------------------- -------------
Sodium 136-145 mEq/L 141 mEq/L
Potassium 3.5-4.5 mEq/L 4.9 mEq/L
Chloride 98-107 mEq/L 100 mEq/L
Osmolality 280-300 mOsm/kg 290 mOsm/kg
Glucose in Fluoride 60-110 mg/dL 76 mg/dL
Creatinine (Jaffé) 0.70-1.20 mg/dL 0.98 mg/dL
CRP \< 5.0 mg/L 4.5 mg/L
Triglycerides \< 150 mg/dL 119 mg/dL
Creatine Kinase \< 190 U/L 142 U/L
Free Triiodothyronine 2.00-4.40 ng/L 3.25 ng/L
Free Thyroxine 9.30-17.00 ng/L 14.12 ng/L
TSH Basal 0.27-4.20 mU/L 1.65 mU/L
Hemoglobin 13.5-17.0 g/dL 14.3 g/dL
Hematocrit 39.5-50.5% 43.4%
Erythrocytes 4.3-5.8 M/uL 5.6 M/uL
Leukocytes 3.90-10.50 K/uL 10.25 K/uL
Platelets 150-370 K/uL 198 K/uL
MCV 80.0-99.0 fL 83.2 fL
MCH 27.0-33.5 pg 28.1 pg
MCHC 31.5-36.0 g/dL 33.4 g/dL
MPV 7.0-12.0 fL 11.6 fL
RDW-CV 11.5-15.0% 13.5%
Quick \> 78% 90%
INR \< 1.25 1.07
aPTT 25.0-38.0 sec 36.1 sec | Malresorptive hydrocephalus |
The Pequod, Nimitz, and Triton are all references to?
A. crewmen aboard the Charles Partlow Sale
B. seafaring men or ships from literature
C. names of scientists who invented food recycling techniques
D. the most palatable strains of algae
| GOURMET By ALLEN KIM LANG [Transcriber's Note: This etext was produced from Galaxy Magazine April 1962. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed.] This was the endless problem of all spaceship cooks: He had to feed the men tomorrow on what they had eaten today! Unable to get out to the ballgame and a long way off from the girls, men on ships think about, talk about, bitch about their food. It's true that Woman remains a topic of thoughtful study, but discussion can never replace practice in an art. Food, on the other hand, is a challenge shipmen face three times a day, so central to their thoughts that a history of sea-faring can be read from a commissary list. In the days when salt-sea sailors were charting islands and spearing seals, for example, the fo'c's'le hands called themselves Lobscousers, celebrating the liquid hash then prominent in the marine menu. The Limey sailor got the name of the anti-scorbutic citrus squeezed into his diet, a fruit known to us mariners of a more sophisticated age only as garnish for our groundside gin-and-tonic. And today we Marsmen are called Slimeheads, honoring in our title the Chlorella and Scenedesmus algae that, by filling up the spaces within, open the road to the larger Space without. Should any groundsman dispute the importance of belly-furniture in history—whether it be exterminating whales, or introducing syphilis to the Fiji Islanders, or settling the Australian littoral with cross-coves from Middlesex and Hampshire—he is referred to the hundred-and-first chapter of Moby Dick , a book spooled in the amusement tanks of all but the smallest spacers. I trust, however, that no Marsman will undertake to review this inventory of refreshment more than a week from groundfall. A catalogue of sides of beef and heads of Leyden cheese and ankers of good Geneva would prove heavy reading for a man condemned to snack on the Chlorella-spawn of cis-Martian space. The Pequod's crew ate wormy biscuit and salt beef. Nimitz's men won their war on canned pork and beans. The Triton made her underwater periplus of Earth with a galley stocked with frozen pizza and concentrated apple-juice. But then, when sailors left the seas for the skies, a decline set in. The first amenity of groundside existence to be abandoned was decent food. The earliest men into the vacuum swallowed protein squeezings from aluminum tubes, and were glad enough to drop back to the groundsman's diet of steak and fried potatoes. Long before I was a boy in Med School, itching to look at black sky through a view-port, galley science had fulfilled the disgusting exordium of Isaiah 36:12, to feed the Slimeheads for breakfast today what was day-before-yesterday's table-scraps and jakes-water. The Ship's Cook, the man who accomplishes the daily miracle of turning offal into eatables, is in many ways the most vital man aboard a spacer. He can make morale or foment a mutiny. His power is paramount. Slimeheads remember the H. M. S. Ajax fiasco, for example, in which a galleyman leveled his Chlorella tanks with heavy water from the ship's shielding. Four officers and twenty-one Other Ranks were rescued from the Ajax in deep space, half dead from deuterium poisoning. We think of the Benjo Maru incident, too, caused by a Ship's Cook who allowed his algaeal staff-of-life to become contaminated with a fast-growing Saccharomycodes yeast. The Japanese vessel staggered to her pad at Piano West after a twenty-week drunk: the alien yeast had got into the stomach of every man aboard, where it fermented each subsequent bite he ate to a superior grade of sake . And for a third footnote to the ancient observation, "God sends food, and the Devil sends cooks," Marsmen will recall what happened aboard my ship the Charles Partlow Sale . The Sale blasted off from Brady Station in the middle of August, due in at Piano West in early May. In no special hurry, we were taking the low-energy route to Mars, a pathway about as long in time as the human period of gestation. Our cargo consisted mostly of Tien-Shen fir seedlings and some tons of an arctic grass-seed—these to be planted in the maria to squeeze out the native blue bugberry vines. We had aboard the Registry minimum of six men and three officers. Ship's Surgeon was myself, Paul Vilanova. Our Captain was Willy Winkelmann, the hardest man in space and very likely the fattest. Ship's Cook was Robert Bailey. Cooking aboard a spacer is a job combining the more frustrating tensions of biochemistry, applied mycology, high-speed farming, dietetics and sewage engineering. It's the Cook's responsibility to see that each man aboard gets each day no less than five pounds of water, two pounds of oxygen, and one-and-a-half pounds of dry food. This isn't just a paragraph from the Spacer Union Contract. It's a statement of the least fuel a man can run on. Twelve tons of water, oxygen, and food would have filled the cargo compartments to bursting, and left a small ship like the C. P. Sale no reason to reach for Mars. By allowing a colony of Chlorella algae to work over our used air, water and other effluvia, though, three tons of metabolites would see us through from Brady Station to Piano West and back. Recycling was the answer. The molecule of carbohydrate, fat, protein or mineral that didn't feed the crew fed the algae. And the algae fed us. All waste was used to fertilize our liquid fields. Even the stubble from our 2,680 shaves and the clippings from our 666 haircuts en route and back would be fed into the Chlorella tanks. Human hair is rich in essential amino acids. The algae—dried by the Cook, bleached with methyl alcohol to kill the smell and make the residue more digestible, disguised and seasoned in a hundred ways—served as a sort of meat-and-potatoes that never quite wore out. Our air and water were equally immortal. Each molecule of oxygen would be conversant with the alveoli of every man aboard by the end of our trip. Every drop of water would have been intimate with the glomeruli of each kidney on the ship before we grounded in. Groundling politicians are right enough when they say that we spacers are a breed apart. We're the one race of men who can't afford the luxury of squeamishness. Though I'm signed aboard as Ship's Surgeon, I seldom lift a knife in space. My employment is more in the nature of TS-card-puncher extraordinary. My duties are to serve as wailing-wall, morale officer, guardian of the medicinal whiskey and frustrator of mutual murder. Generally the man aboard who'd serve as the most popular murder-victim is the Cook. This trip, the-man-you-love-to-hate was our Captain. If the Cook hadn't problems enough with the chemical and psychic duties of his office, Winkelmann supplied the want. Captain Willy Winkelmann was the sort of man who, if he had to go into space at all, had best do so alone. If the Prussians had a Marine Corps, Winkelmann would have done splendidly as Drill Instructor for their boot camp. His heart was a chip of helium ice, his voice dripped sarcastic acid. The planet Earth was hardly large enough to accommodate a wart as annoying as Willy Winkelmann. Cheek-by-jowl every day in a nacelle the size of a Pullman car, our Captain quickly established himself as a major social hemorrhoid. The Captain's particular patsy was, of course, young Bailey the Cook. It was Winkelmann who saw humorous possibilities in the entry, "Bailey, Robert," on Ship's Articles. He at once renamed our unfortunate shipmate "Belly-Robber." It was Winkelmann who discussed haut cuisine and the properties of the nobler wines while we munched our algaeburgers and sipped coffee that tasted of utility water. And it was Captain Willy Winkelmann who never referred to the ship's head by any other name than The Kitchen Cabinet. Bailey tried to feed us by groundside standards. He hid the taste of synthetic methionine—an essential amino acid not synthesized by Chlorella—by seasoning our algaeal repasts with pinches of oregano and thyme. He tinted the pale-green dollops of pressed Chlorella pink, textured the mass to the consistency of hamburger and toasted the slabs to a delicate brown in a forlorn attempt to make mock-meat. For dessert, he served a fudge compounded from the dextrose-paste of the carbohydrate recycler. The crew thanked him. The Captain did not. "Belly-Robber," he said, his tone icy as winter wind off the North Sea, "you had best cycle this mess through the tanks again. There is a pun in my home country: Mensch ist was er isst. It means, you are what you eat. I think you are impertinent to suggest I should become this Schweinerei you are feeding me." Captain Winkelmann blotted his chin with his napkin, heaved his bulk up from the table, and climbed up the ladder from the dining-cubby. "Doc, do you like Winkelmann?" the Cook asked me. "Not much," I said. "I suspect that the finest gift our Captain can give his mother is to be absent from her on Mother's Day. But we've got to live with him. He's a good man at driving a ship." "I wish he'd leave off driving this Cook," Bailey said. "The fat swine!" "His plumpness is an unwitting tribute to your cooking, Bailey," I said. "He eats well. We all do. I've dined aboard a lot of spacers in my time, and I'll testify that you set a table second to none." Bailey took a handful of dried Chlorella from a bin and fingered it. It was green, smelled of swamp, and looked appetizing as a bedsore. "This is what I have to work with," he said. He tossed the stuff back into its bin. "In Ohio, which is my home country, in the presence of ladies, we'd call such garbage Horse-Leavings." "You'll never make Winkelmann happy," I said. "Even the simultaneous death of all other human beings could hardly make him smile. Keep up the good work, though, and you'll keep our Captain fat." Bailey nodded from his one-man cloud of gloom. I got a bottle of rye from Medical Stores and offered him a therapeutic draught. The Cook waved my gift aside. "Not now, Doc," he said. "I'm thinking about tomorrow's menu." The product of Bailey's cerebrations was on the mess table at noon the next day. We were each served an individual head of lettuce, dressed with something very like vinegar and oil, spiced with tiny leaves of burnet. How Bailey had constructed those synthetic lettuces I can only guess: the hours spent preparing a green Chlorella paste, rolling and drying and shaping each artificial leaf, the fitting together of nine heads like crisp, three-dimensional jigsaw puzzles. The pièce de résistance was again a "hamburger steak;" but this time the algaeal mass that made it up was buried in a rich, meaty gravy that was only faintly green. The essence-of-steak used in these Chlorella cutlets had been sprinkled with a lavish hand. Garlic was richly in evidence. "It's so tender," the radioman joked, "that I can hardly believe it's really steak." Bailey stared across the dining-cubby toward Winkelmann, silently imploring the Captain's ratification of his masterpiece. The big man's pink cheeks bulged and jumped with his chewing. He swallowed. "Belly-Robber," Winkelmann said, "I had almost rather you served me this pond-scum raw than have it all mucked-up with synthetic onions and cycler-salt." "You seem able enough to choke down Bailey's chow, Captain," I said. I gazed at Winkelmann's form, bulbous from a lifetime of surfeit feeding. "Yes, I eat it," the Captain said, taking and talking through another bite. "But I eat only as a man in the desert will eat worms and grasshoppers, to stay alive." "Sir, what in heaven's name do you expect from me?" Bailey pleaded. "Only good food," Winkelmann mumbled through his mouthful of disguised algae. He tapped his head with a finger. "This—the brain that guides the ship—cannot be coaxed to work on hog-slop. You understand me, Belly-Robber?" Bailey, his hands fisted at his sides, nodded. "Yes, sir. But I really don't know what I can do to please you." "You are a spacer and a Ship's Cook, not a suburban Hausfrau with the vapors," Winkelmann said. "I do not expect from you hysterics, tantrums or weeping. Only—can you understand this, so simple?—food that will keep my belly content and my brain alive." "Yes, sir," Bailey said, his face a picture of that offense the British term Dumb Insolence. Winkelmann got up and climbed the ladder to the pilot-cubicle. I followed him. "Captain," I said, "you're driving Bailey too hard. You're asking him to make bricks without straw." Winkelmann regarded me with his pale-blue stare. "You think, Doctor, that my cruelty to the Belly-Robber is the biliousness of a middle-aged man?" "Frankly, I can't understand your attitude at all," I said. "You accuse me of driving a man to make bricks without straw," Winkelmann said. "Very well, Doctor. It is my belief that if the Pharaoh's taskmaster had had my firmness of purpose, the Children of Israel would have made bricks with stubble. Necessity, Doctor, is the mother of invention. I am Bailey's necessity. My unkindnesses make him uncomfortable, I doubt that not. But I am forcing him to experiment, to improvise, to widen the horizons of his ingenuity. He will learn somehow to bring good food from Chlorella tanks." "You're driving him too hard, Sir," I said. "He'll crack." "Bailey will have some fifty thousand dollars' salary waiting when we ground at Brady Station," Captain Winkelmann said. "So much money buys many discomforts. That will be all, Doctor Vilanova." "Crew morale on the ship...." I began. "That will be all, Doctor Vilanova," Captain Winkelmann repeated. Bailey grew more silent as we threaded our way along the elliptical path to Mars. Each meal he prepared was a fresh attempt to propitiate the appetite of our splenetic Captain. Each such offering was condemned by that heartless man. Bailey began to try avoiding the Captain at mealtimes, but was frustrated by Winkelmann's orders. "Convey my compliments to the Chef, please," the Captain would instruct one of the crew, "and ask him to step down here a moment." And the Cook would cheerlessly appear in the dining-cubby, to have his culinary genius acidly called in question again. I myself do not doubt that Bailey was the finest Cook ever to go into Hohmann orbit. His every meal established a higher benchmark in brilliant galleymanship. We were served, for instance, an ersatz hot turkey supreme. The cheese-sauce was almost believable, the Chlorella turkey-flesh was white and tender. Bailey served with this delicacy a grainy and delicious "cornbread," and had extracted from his algae a lipid butter-substitute that soaked into the hot "bread" with a genuinely dairy smell. "Splendid, Bailey," I said. "We are not amused," said Captain Winkelmann, accepting a second helping of the pseudo-turkey. "You are improving, Belly-Robber, but only arithmetically. Your first efforts were so hideous as to require a geometric progression of improving excellence to raise them to mere edibility. By the time we are halfway 'round the Sun, I trust you will have learned to cook with the competence of a freshman Home Economics student. That will be all, Bailey." The crew and my fellow-officers were amused by Winkelmann's riding of Bailey; they were in addition gratified that the battle between their Captain and their Cook served to feed them so well. Most spacers embark on an outward voyage somewhat plump, having eaten enough on their last few days aground to smuggle several hundred calories of fat and many memories of good food aboard with them. This trip, none of the men had lost weight during the first four months in space. Winkelmann, indeed, seemed to have gained. His uniform was taut over his plump backside, and he puffed a bit up the ladders. I was considering suggesting to our Captain that he curtail his diet for reasons of health, a bit of advice that would have stood unique in the annals of space medicine, when Winkelmann produced his supreme insult to our Cook. Each man aboard a spacer is allowed ten kilograms of personal effects besides his uniforms, these being considered Ship's Furnishing. As his rank and responsibility merit, the Captain is allowed double this ration. He may thus bring aboard with him some forty-five pounds of books, playing-cards, knitting-wool, whiskey or what have you to help him while away the hours between the planets. Bailey, I knew for a fact, had used up his weight-allowance in bringing aboard a case of spices: marjoram and mint, costmary, file powder, basil and allspice, and a dozen others. Captain Winkelmann was not a reader, and had brought no books. Cards interested him not at all, as card-playing implies a sociability alien to his nature. He never drank aboard ship. I had supposed that he'd exercised his option of returning his personal-effects weight allowance to the owners for the consideration of one hundred dollars a kilogram. To collect the maximum allowance, spacers have been known to come aboard their ship mother-naked. But this was not the case with Winkelmann. His personal-effects baggage, an unlabeled cardboard box, appeared under the table at noon mess some hundred days out from Piano West. Winkelmann rested his feet on the mysterious box as he sat to eat. "What disgusting form does the ship's garbage appear in today, Belly-Robber?" he asked the Cook. Bailey frowned, but kept his temper, an asceticism in which by now he'd had much practice. "I've been working on the problem of steak, Sir," he said. "I think I've whipped the taste; what was left was to get the texture steak-like. Do you understand, Sir?" "I understand," Winkelmann growled. "You intend that your latest mess should feel like steak to the mouth, and not like baby-food. Right?" "Yes, Sir," Bailey said. "Well, I squeezed the steak-substrate—Chlorella, of course, with all sorts of special seasonings—through a sieve, and blanched the strands in hot algaeal oil. Then I chopped those strands to bits and rolled them out. Voila! I had something very close in texture to the muscle-fibers of genuine meat." "Remarkable, Bailey," I said. "It rather throws me off my appetite to hear how you muddle about with our food," the Captain said, his jowls settling into an expression of distaste. "It's quite all right to eat lobster, for example, but I never cared to see the ugly beast boiled before my eyes. Detail spoils the meal." Bailey lifted the cover off the electric warming-pan at the center of the table and tenderly lifted a small "steak" onto each of our plates. "Try it," he urged the Captain. Captain Winkelmann sliced off a corner of his algaeal steak. The color was an excellent medium-rare, the odor was the rich smell of fresh-broiled beef. Winkelmann bit down, chewed, swallowed. "Not too bad, Belly-Robber," he said, nodding. Bailey grinned and bobbed his head, his hands folded before him in an ecstasy of pleasure. A kind word from the Captain bettered the ruffles-and-flourishes of a more reasonable man. "But it still needs something ... something," Winkelmann went on, slicing off another portion of the tasty Chlorella. "Aha! I have it!" "Yes, Sir?" Bailey asked. "This, Belly-Robber!" Winkelmann reached beneath the mess-table and ripped open his cardboard carton. He brought out a bottle and unscrewed the cap. "Ketchup," he said, splattering the red juice over Bailey's masterpiece. "The scarlet burial-shroud for the failures of Cooks." Lifting a hunk of the "steak," streaming ketchup, to his mouth, Winkelmann chewed. "Just the thing," he smiled. "Damn you!" Bailey shouted. Winkelmann's smile flicked off, and his blue eyes pierced the Cook. "... Sir," Bailey added. "That's better," Winkelmann said, and took another bite. He said meditatively, "Used with caution, and only by myself, I believe I have sufficient ketchup here to see me through to Mars. Please keep a bottle on the table for all my future meals, Belly-Robber." "But, Sir...." Bailey began. "You must realize, Belly-Robber, that a dyspeptic Captain is a threat to the welfare of his ship. Were I to continue eating your surrealistic slops for another hundred days, without the small consolation of this sauce I had the foresight to bring with me, I'd likely be in no condition to jet us safely down to the Piano West pad. Do you understand, Belly-Robber?" he demanded. "I understand that you're an ungrateful, impossible, square-headed, slave-driving...." "Watch your noun," Winkelmann cautioned the Cook. "Your adjectives are insubordinate; your noun might prove mutinous." "Captain, you've gone too far," I said. Bailey, his fists knotted, was scarlet, his chest heaving with emotion. "Doctor, I must point out to you that it ill behooves the Ship's Surgeon to side with the Cook against the Captain," Winkelmann said. "Sir, Bailey has tried hard to please you," I said. "The other officers and the men have been more than satisfied with his work." "That only suggests atrophy of their taste buds," Winkelmann said. "Doctor, you are excused. As are you, Belly-Robber," he added. Bailey and I climbed from the mess compartment together. I steered him to my quarters, where the medical supplies were stored. He sat on my bunk and exploded into weeping, banging his fists against the metal bulkhead. "You'll have that drink now," I said. "No, dammit!" he shouted. "Orders," I said. I poured us each some fifty cc's of rye. "This is therapy, Bailey," I told him. He poured the fiery stuff down his throat like water and silently held out his glass for a second. I provided it. After a few minutes Bailey's sobbing ceased. "Sorry, Doc," he said. "You've taken more pressure than most men would," I said. "Nothing to be ashamed of." "He's crazy. What sane man would expect me to dip Wiener schnitzel and sauerkraut and Backhahndl nach suddeutscher Art out of an algae tank? I've got nothing but microscopic weeds to cook for him! Worn-out molecules reclaimed from the head; packaged amino acid additives. And he expects meals that would take the blue ribbon at the annual banquet of the Friends of Escoffier!" "Yours is an ancient plaint, Bailey," I said. "You've worked your fingers to the bone, slaving over a hot stove, and you're not appreciated. But you're not married to Winkelmann, remember. A year from now you'll be home in Ohio, fifty grand richer, set to start that restaurant of yours and forget about our fat Flying Dutchman." "I hate him," Bailey said with the simplicity of true emotion. He reached for the bottle. I let him have it. Sometimes alcohol can be an apt confederate of vis medicatrix naturae , the healing power of nature. Half an hour later I strapped Bailey into his bunk to sleep it off. That therapeutic drunk seemed to be just what he'd needed. For morning mess the next day we had a broth remarkable in horribleness, a pottage or boiled Chlorella vulgaris that looked and tasted like the vomit of some bottom-feeding sea-beast. Bailey, red-eyed and a-tremble, made no apology, and stared at Winkelmann as though daring him to comment. The Captain lifted a spoonful of the disgusting stuff to his lips, smacked and said, "Belly-Robber, you're improving a little at last." Bailey nodded and smiled. "Thank you, Sir," he said. I smiled, too. Bailey had conquered himself. His psychic defenses were now strong enough to withstand the Captain's fiercest assaults of irony. Our food would likely be bad the rest of this trip, but that was a price I was willing to pay for seeing destroyed the Willy Winkelmann theory of forcing a Cook to make bricks without straw. The Captain had pushed too hard. He'd need that ketchup for the meals to come, I thought. Noon mess was nearly as awful as breakfast had been. The coffee tasted of salt, and went largely undrunk. The men in the mess compartment were vehement in their protests, blaming the Captain, in his absence, for the decline in culinary standards. Bailey seemed not to care. He served the algaeburgers with half a mind, and hurried back into his galley oblivious of the taunts of his crewmates. There being only three seats in the Sale's mess compartment, we ate our meals in three shifts. That evening, going down the ladder to supper, my nose was met with a spine-tingling barbecue tang, a smell to make a man think of gray charcoal glowing in a picnic brazier, of cicadas chirping and green grass underfoot, of the pop and hiss of canned beer being church-keyed. "He's done it, Doc!" one of the first-shift diners said. "It actually tastes of food!" "Then he's beat the Captain at his game," I said. "The Dutchman won't want to mess ketchup on these steaks," the crewman said. I sat, unfolded my napkin, and looked with hope to the electric warming-pan at the center of the table. Bailey served the three of us with the small "steaks." Each contained about a pound of dried Chlorella, I judged, teasing mine with my fork. But they were drenched in a gravy rich as the stuff grandma used to make in her black iron skillet, peppery and seasoned with courageous bits of garlic. I cut a bit from my steak and chewed it. Too tender, of course; there are limits to art. But the pond-scum taste was gone. Bailey appeared in the galley door. I gestured for him to join me. "You've done it, Bailey," I said. "Every Slimehead in orbit will thank you for this. This is actually good ." "Thanks, Doc," Bailey said. I smiled and took another bite. "You may not realize it, Bailey; but this is a victory for the Captain, too. He drove you to this triumph; you couldn't have done it without him." "You mean he was just whipping me on, trying to make me do better?" Bailey asked. "He was driving you to do the impossible," I said; "and you did it. Our Captain may be a hard man, Bailey; but he did know how to coax maximum performance out of his Ship's Cook." Bailey stood up. "Do you like Captain Winkelmann, Doctor?" he asked. I thought about his question a moment. Winkelmann was good at his job. He persuaded his men by foul means, true; but it was all for the good of the ship and his crew. "Do I like Captain Winkelmann?" I asked, spearing another piece of my artificial steak. "Bailey, I'm afraid I'll have to admit that I do." Bailey smiled and lifted a second steak from the warming-pan onto my plate. "Then have another piece," he said. | B. seafaring men or ships from literature |
Referring to the passage’s title, who was the “Monster Maker”?
A. Click
B. Human imagination
C. Gunther
D. Irish
| The Monster Maker By RAY BRADBURY "Get Gunther," the official orders read. It was to laugh! For Click and Irish were marooned on the pirate's asteroid—their only weapons a single gun and a news-reel camera. [Transcriber's Note: This etext was produced from Planet Stories Spring 1944. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed.] Suddenly, it was there. There wasn't time to blink or speak or get scared. Click Hathaway's camera was loaded and he stood there listening to it rack-spin film between his fingers, and he knew he was getting a damned sweet picture of everything that was happening. The picture of Marnagan hunched huge over the control-console, wrenching levers, jamming studs with freckled fists. And out in the dark of the fore-part there was space and a star-sprinkling and this meteor coming like blazing fury. Click Hathaway felt the ship move under him like a sensitive animal's skin. And then the meteor hit. It made a spiked fist and knocked the rear-jets flat, and the ship spun like a cosmic merry-go-round. There was plenty of noise. Too damned much. Hathaway only knew he was picked up and hurled against a lever-bank, and that Marnagan wasn't long in following, swearing loud words. Click remembered hanging on to his camera and gritting to keep holding it. What a sweet shot that had been of the meteor! A sweeter one still of Marnagan beating hell out of the controls and keeping his words to himself until just now. It got quiet. It got so quiet you could almost hear the asteroids rushing up, cold, blue and hard. You could hear your heart kicking a tom-tom between your sick stomach and your empty lungs. Stars, asteroids revolved. Click grabbed Marnagan because he was the nearest thing, and held on. You came hunting for a space-raider and you ended up cradled in a slab-sized Irishman's arms, diving at a hunk of metal death. What a fade-out! "Irish!" he heard himself say. "Is this IT?" "Is this what ?" yelled Marnagan inside his helmet. "Is this where the Big Producer yells CUT!?" Marnagan fumed. "I'll die when I'm damned good and ready. And when I'm ready I'll inform you and you can picture me profile for Cosmic Films!" They both waited, thrust against the shipside and held by a hand of gravity; listening to each other's breathing hard in the earphones. The ship struck, once. Bouncing, it struck again. It turned end over and stopped. Hathaway felt himself grabbed; he and Marnagan rattled around—human dice in a croupier's cup. The shell of the ship burst, air and energy flung out. Hathaway screamed the air out of his lungs, but his brain was thinking quick crazy, unimportant things. The best scenes in life never reach film, or an audience. Like this one, dammit! Like this one! His brain spun, racketing like the instantaneous, flicking motions of his camera. Silence came and engulfed all the noise, ate it up and swallowed it. Hathaway shook his head, instinctively grabbed at the camera locked to his mid-belt. There was nothing but stars, twisted wreckage, cold that pierced through his vac-suit, and silence. He wriggled out of the wreckage into that silence. He didn't know what he was doing until he found the camera in his fingers as if it had grown there when he was born. He stood there, thinking "Well, I'll at least have a few good scenes on film. I'll—" A hunk of metal teetered, fell with a crash. Marnagan elevated seven feet of bellowing manhood from the wreck. "Hold it!" cracked Hathaway's high voice. Marnagan froze. The camera whirred. "Low angle shot; Interplanetary Patrolman emerges unscathed from asteroid crackup. Swell stuff. I'll get a raise for this!" "From the toe of me boot!" snarled Marnagan brusquely. Oxen shoulders flexed inside his vac-suit. "I might've died in there, and you nursin' that film-contraption!" Hathaway felt funny inside, suddenly. "I never thought of that. Marnagan die? I just took it for granted you'd come through. You always have. Funny, but you don't think about dying. You try not to." Hathaway stared at his gloved hand, but the gloving was so thick and heavy he couldn't tell if it was shaking. Muscles in his bony face went down, pale. "Where are we?" "A million miles from nobody." They stood in the middle of a pocked, time-eroded meteor plain that stretched off, dipping down into silent indigo and a rash of stars. Overhead, the sun poised; black and stars all around it, making it look sick. "If we walk in opposite directions, Click Hathaway, we'd be shaking hands the other side of this rock in two hours." Marnagan shook his mop of dusty red hair. "And I promised the boys at Luna Base this time I'd capture that Gunther lad!" His voice stopped and the silence spoke. Hathaway felt his heart pumping slow, hot pumps of blood. "I checked my oxygen, Irish. Sixty minutes of breathing left." The silence punctuated that sentence, too. Upon the sharp meteoric rocks Hathaway saw the tangled insides of the radio, the food supply mashed and scattered. They were lucky to have escaped. Or was suffocation a better death...? Sixty minutes. They stood and looked at one another. "Damn that meteor!" said Marnagan, hotly. Hathaway got hold of an idea; remembering something. He said it out: "Somebody tossed that meteor, Irish. I took a picture of it, looked it right in the eye when it rolled at us, and it was poker-hot. Space-meteors are never hot and glowing. If it's proof you want, I've got it here, on film." Marnagan winced his freckled square of face. "It's not proof we need now, Click. Oxygen. And then food . And then some way back to Earth." Hathaway went on saying his thoughts: "This is Gunther's work. He's here somewhere, probably laughing his guts out at the job he did us. Oh, God, this would make great news-release stuff if we ever get back to Earth. I.P.'s Irish Marnagan, temporarily indisposed by a pirate whose dirty face has never been seen, Gunther by name, finally wins through to a triumphant finish. Photographed on the spot, in color, by yours truly, Click Hathaway. Cosmic Films, please notice." They started walking, fast, over the pocked, rubbled plain toward a bony ridge of metal. They kept their eyes wide and awake. There wasn't much to see, but it was better than standing still, waiting. Marnagan said, "We're working on margin, and we got nothin' to sweat with except your suspicions about this not being an accident. We got fifty minutes to prove you're right. After that—right or wrong—you'll be Cosmic Films prettiest unmoving, unbreathin' genius. But talk all you like, Click. It's times like this when we all need words, any words, on our tongues. You got your camera and your scoop. Talk about it. As for me—" he twisted his glossy red face. "Keeping alive is me hobby. And this sort of two-bit death I did not order." Click nodded. "Gunther knows how you'd hate dying this way, Irish. It's irony clean through. That's probably why he planned the meteor and the crash this way." Marnagan said nothing, but his thick lips went down at the corners, far down, and the green eyes blazed. They stopped, together. "Oops!" Click said. "Hey!" Marnagan blinked. "Did you feel that ?" Hathaway's body felt feathery, light as a whisper, boneless and limbless, suddenly. "Irish! We lost weight, coming over that ridge!" They ran back. "Let's try it again." They tried it. They scowled at each other. The same thing happened. "Gravity should not act this way, Click." "Are you telling me? It's man-made. Better than that—it's Gunther! No wonder we fell so fast—we were dragged down by a super-gravity set-up! Gunther'd do anything to—did I say anything ?" Hathaway leaped backward in reaction. His eyes widened and his hand came up, jabbing. Over a hill-ridge swarmed a brew of unbelievable horrors. Progeny from Frankenstein's ARK. Immense crimson beasts with numerous legs and gnashing mandibles, brown-black creatures, some tubular and fat, others like thin white poisonous whips slashing along in the air. Fangs caught starlight white on them. Hathaway yelled and ran, Marnagan at his heels, lumbering. Sweat broke cold on his body. The immense things rolled, slithered and squirmed after him. A blast of light. Marnagan, firing his proton-gun. Then, in Click's ears, the Irishman's incredulous bellow. The gun didn't hurt the creatures at all. "Irish!" Hathaway flung himself over the ridge, slid down an incline toward the mouth a small cave. "This way, fella!" Hathaway made it first, Marnagan bellowing just behind him. "They're too big; they can't get us in here!" Click's voice gasped it out, as Marnagan squeezed his two-hundred-fifty pounds beside him. Instinctively, Hathaway added, "Asteroid monsters! My camera! What a scene!" "Damn your damn camera!" yelled Marnagan. "They might come in!" "Use your gun." "They got impervious hides. No use. Gahh! And that was a pretty chase, eh, Click?" "Yeah. Sure. You enjoyed it, every moment of it." "I did that." Irish grinned, showing white uneven teeth. "Now, what will we be doing with these uninvited guests at our door?" "Let me think—" "Lots of time, little man. Forty more minutes of air, to be exact." They sat, staring at the monsters for about a minute. Hathaway felt funny about something; didn't know what. Something about these monsters and Gunther and— "Which one will you be having?" asked Irish, casually. "A red one or a blue one?" Hathaway laughed nervously. "A pink one with yellow ruffles—Good God, now you've got me doing it. Joking in the face of death." "Me father taught me; keep laughing and you'll have Irish luck." That didn't please the photographer. "I'm an Anglo-Swede," he pointed out. Marnagan shifted uneasily. "Here, now. You're doing nothing but sitting, looking like a little boy locked in a bedroom closet, so take me a profile shot of the beasties and myself." Hathaway petted his camera reluctantly. "What in hell's the use? All this swell film shot. Nobody'll ever see it." "Then," retorted Marnagan, "we'll develop it for our own benefit; while waitin' for the U.S. Cavalry to come riding over the hill to our rescue!" Hathaway snorted. "U.S. Cavalry." Marnagan raised his proton-gun dramatically. "Snap me this pose," he said. "I paid your salary to trot along, photographing, we hoped, my capture of Gunther, now the least you can do is record peace negotiations betwixt me and these pixies." Marnagan wasn't fooling anybody. Hathaway knew the superficial palaver for nothing but a covering over the fast, furious thinking running around in that red-cropped skull. Hathaway played the palaver, too, but his mind was whirring faster than his camera as he spun a picture of Marnagan standing there with a useless gun pointed at the animals. Montage. Marnagan sitting, chatting at the monsters. Marnagan smiling for the camera. Marnagan in profile. Marnagan looking grim, without much effort, for the camera. And then, a closeup of the thrashing death wall that holed them in. Click took them all, those shots, not saying anything. Nobody fooled nobody with this act. Death was near and they had sweaty faces, dry mouths and frozen guts. When Click finished filming, Irish sat down to save oxygen, and used it up arguing about Gunther. Click came back at him: "Gunther drew us down here, sure as Ceres! That gravity change we felt back on that ridge, Irish; that proves it. Gunther's short on men. So, what's he do; he builds an asteroid-base, and drags ships down. Space war isn't perfect yet, guns don't prime true in space, trajectory is lousy over long distances. So what's the best weapon, which dispenses with losing valuable, rare ships and a small bunch of men? Super-gravity and a couple of well-tossed meteors. Saves all around. It's a good front, this damned iron pebble. From it, Gunther strikes unseen; ships simply crash, that's all. A subtle hand, with all aces." Marnagan rumbled. "Where is the dirty son, then!" "He didn't have to appear, Irish. He sent—them." Hathaway nodded at the beasts. "People crashing here die from air-lack, no food, or from wounds caused at the crackup. If they survive all that—the animals tend to them. It all looks like Nature was responsible. See how subtle his attack is? Looks like accidental death instead of murder, if the Patrol happens to land and finds us. No reason for undue investigation, then." "I don't see no Base around." Click shrugged. "Still doubt it? Okay. Look." He tapped his camera and a spool popped out onto his gloved palm. Holding it up, he stripped it out to its full twenty inch length, held it to the light while it developed, smiling. It was one of his best inventions. Self-developing film. The first light struck film-surface, destroyed one chemical, leaving imprints; the second exposure simply hardened, secured the impressions. Quick stuff. Inserting the film-tongue into a micro-viewer in the camera's base, Click handed the whole thing over. "Look." Marnagan put the viewer up against the helmet glass, squinted. "Ah, Click. Now, now. This is one lousy film you invented." "Huh?" "It's a strange process'll develop my picture and ignore the asteroid monsters complete." "What!" Hathaway grabbed the camera, gasped, squinted, and gasped again: Pictures in montage; Marnagan sitting down, chatting conversationally with nothing ; Marnagan shooting his gun at nothing ; Marnagan pretending to be happy in front of nothing . Then, closeup—of—NOTHING! The monsters had failed to image the film. Marnagan was there, his hair like a red banner, his freckled face with the blue eyes bright in it. Maybe— Hathaway said it, loud: "Irish! Irish! I think I see a way out of this mess! Here—" He elucidated it over and over again to the Patrolman. About the film, the beasts, and how the film couldn't be wrong. If the film said the monsters weren't there, they weren't there. "Yeah," said Marnagan. "But step outside this cave—" "If my theory is correct I'll do it, unafraid," said Click. Marnagan scowled. "You sure them beasts don't radiate ultra-violet or infra-red or something that won't come out on film?" "Nuts! Any color we see, the camera sees. We've been fooled." "Hey, where you going?" Marnagan blocked Hathaway as the smaller man tried pushing past him. "Get out of the way," said Hathaway. Marnagan put his big fists on his hips. "If anyone is going anywhere, it'll be me does the going." "I can't let you do that, Irish." "Why not?" "You'd be going on my say-so." "Ain't your say-so good enough for me?" "Yes. Sure. Of course. I guess—" "If you say them animals ain't there, that's all I need. Now, stand aside, you film-developing flea, and let an Irishman settle their bones." He took an unnecessary hitch in trousers that didn't exist except under an inch of porous metal plate. "Your express purpose on this voyage, Hathaway, is taking films to be used by the Patrol later for teaching Junior Patrolmen how to act in tough spots. First-hand education. Poke another spool of film in that contraption and give me profile a scan. This is lesson number seven: Daniel Walks Into The Lion's Den." "Irish, I—" "Shut up and load up." Hathaway nervously loaded the film-slot, raised it. "Ready, Click?" "I—I guess so," said Hathaway. "And remember, think it hard, Irish. Think it hard. There aren't any animals—" "Keep me in focus, lad." "All the way, Irish." "What do they say...? Oh, yeah. Action. Lights. Camera!" Marnagan held his gun out in front of him and still smiling took one, two, three, four steps out into the outside world. The monsters were waiting for him at the fifth step. Marnagan kept walking. Right out into the middle of them.... That was the sweetest shot Hathaway ever took. Marnagan and the monsters! Only now it was only Marnagan. No more monsters. Marnagan smiled a smile broader than his shoulders. "Hey, Click, look at me! I'm in one piece. Why, hell, the damned things turned tail and ran away!" "Ran, hell!" cried Hathaway, rushing out, his face flushed and animated. "They just plain vanished. They were only imaginative figments!" "And to think we let them hole us in that way, Click Hathaway, you coward!" "Smile when you say that, Irish." "Sure, and ain't I always smilin'? Ah, Click boy, are them tears in your sweet grey eyes?" "Damn," swore the photographer, embarrassedly. "Why don't they put window-wipers in these helmets?" "I'll take it up with the Board, lad." "Forget it. I was so blamed glad to see your homely carcass in one hunk, I couldn't help—Look, now, about Gunther. Those animals are part of his set-up. Explorers who land here inadvertently, are chased back into their ships, forced to take off. Tourists and the like. Nothing suspicious about animals. And if the tourists don't leave, the animals kill them." "Shaw, now. Those animals can't kill." "Think not, Mr. Marnagan? As long as we believed in them they could have frightened us to death, forced us, maybe, to commit suicide. If that isn't being dangerous—" The Irishman whistled. "But, we've got to move , Irish. We've got twenty minutes of oxygen. In that time we've got to trace those monsters to their source, Gunther's Base, fight our way in, and get fresh oxy-cannisters." Click attached his camera to his mid-belt. "Gunther probably thinks we're dead by now. Everyone else's been fooled by his playmates; they never had a chance to disbelieve them." "If it hadn't been for you taking them pictures, Click—" "Coupled with your damned stubborn attitude about the accident—" Click stopped and felt his insides turning to water. He shook his head and felt a film slip down over his eyes. He spread his legs out to steady himself, and swayed. "I—I don't think my oxygen is as full as yours. This excitement had me double-breathing and I feel sick." Marnagan's homely face grimaced in sympathy. "Hold tight, Click. The guy that invented these fish-bowls didn't provide for a sick stomach." "Hold tight, hell, let's move. We've got to find where those animals came from! And the only way to do that is to get the animals to come back!" "Come back? How?" "They're waiting, just outside the aura of our thoughts, and if we believe in them again, they'll return." Marnagan didn't like it. "Won't—won't they kill us—if they come—if we believe in 'em?" Hathaway shook a head that was tons heavy and weary. "Not if we believe in them to a certain point . Psychologically they can both be seen and felt. We only want to see them coming at us again." " Do we, now?" "With twenty minutes left, maybe less—" "All right, Click, let's bring 'em back. How do we do it?" Hathaway fought against the mist in his eyes. "Just think—I will see the monsters again. I will see them again and I will not feel them. Think it over and over." Marnagan's hulk stirred uneasily. "And—what if I forget to remember all that? What if I get excited...?" Hathaway didn't answer. But his eyes told the story by just looking at Irish. Marnagan cursed. "All right, lad. Let's have at it!" The monsters returned. A soundless deluge of them, pouring over the rubbled horizon, swarming in malevolent anticipation about the two men. "This way, Irish. They come from this way! There's a focal point, a sending station for these telepathic brutes. Come on!" Hathaway sludged into the pressing tide of color, mouths, contorted faces, silvery fat bodies misting as he plowed through them. Marnagan was making good progress ahead of Hathaway. But he stopped and raised his gun and made quick moves with it. "Click! This one here! It's real!" He fell back and something struck him down. His immense frame slammed against rock, noiselessly. Hathaway darted forward, flung his body over Marnagan's, covered the helmet glass with his hands, shouting: "Marnagan! Get a grip, dammit! It's not real—don't let it force into your mind! It's not real, I tell you!" "Click—" Marnagan's face was a bitter, tortured movement behind glass. "Click—" He was fighting hard. "I—I—sure now. Sure—" He smiled. "It—it's only a shanty fake!" "Keep saying it, Irish. Keep it up." Marnagan's thick lips opened. "It's only a fake," he said. And then, irritated, "Get the hell off me, Hathaway. Let me up to my feet!" Hathaway got up, shakily. The air in his helmet smelled stale, and little bubbles danced in his eyes. "Irish, you forget the monsters. Let me handle them, I know how. They might fool you again, you might forget." Marnagan showed his teeth. "Gah! Let a flea have all the fun? And besides, Click, I like to look at them. They're pretty." The outpour of animals came from a low lying mound a mile farther on. Evidently the telepathic source lay there. They approached it warily. "We'll be taking our chances on guard," hissed Irish. "I'll go ahead, draw their attention, maybe get captured. Then, you show up with your gun...." "I haven't got one." "We'll chance it, then. You stick here until I see what's ahead. They probably got scanners out. Let them see me—" And before Hathaway could object, Marnagan walked off. He walked about five hundred yards, bent down, applied his fingers to something, heaved up, and there was a door opening in the rock. His voice came back across the distance, into Click's earphones. "A door, an air-lock, Click. A tunnel leading down inside!" Then, Marnagan dropped into the tunnel, disappearing. Click heard the thud of his feet hitting the metal flooring. Click sucked in his breath, hard and fast. "All right, put 'em up!" a new harsh voice cried over a different radio. One of Gunther's guards. Three shots sizzled out, and Marnagan bellowed. The strange harsh voice said, "That's better. Don't try and pick that gun up now. Oh, so it's you. I thought Gunther had finished you off. How'd you get past the animals?" Click started running. He switched off his sending audio, kept his receiving on. Marnagan, weaponless. One guard. Click gasped. Things were getting dark. Had to have air. Air. Air. He ran and kept running and listening to Marnagan's lying voice: "I tied them pink elephants of Gunther's in neat alphabetical bundles and stacked them up to dry, ya louse!" Marnagan said. "But, damn you, they killed my partner before he had a chance!" The guard laughed. The air-lock door was still wide open when Click reached it, his head swimming darkly, his lungs crammed with pain-fire and hell-rockets. He let himself down in, quiet and soft. He didn't have a weapon. He didn't have a weapon. Oh, damn, damn! A tunnel curved, ending in light, and two men silhouetted in that yellow glare. Marnagan, backed against a wall, his helmet cracked, air hissing slowly out of it, his face turning blue. And the guard, a proton gun extended stiffly before him, also in a vac-suit. The guard had his profile toward Hathaway, his lips twisting: "I think I'll let you stand right there and die," he said quietly. "That what Gunther wanted, anway. A nice sordid death." Hathaway took three strides, his hands out in front of him. "Don't move!" he snapped. "I've got a weapon stronger than yours. One twitch and I'll blast you and the whole damned wall out from behind you! Freeze!" The guard whirled. He widened his sharp eyes, and reluctantly, dropped his gun to the floor. "Get his gun, Irish." Marnagan made as if to move, crumpled clumsily forward. Hathaway ran in, snatched up the gun, smirked at the guard. "Thanks for posing," he said. "That shot will go down in film history for candid acting." "What!" "Ah: ah! Keep your place. I've got a real gun now. Where's the door leading into the Base?" The guard moved his head sullenly over his left shoulder. Click was afraid he would show his weak dizziness. He needed air. "Okay. Drag Marnagan with you, open the door and we'll have air. Double time! Double!" Ten minutes later, Marnagan and Hathaway, fresh tanks of oxygen on their backs, Marnagan in a fresh bulger and helmet, trussed the guard, hid him in a huge trash receptacle. "Where he belongs," observed Irish tersely. They found themselves in a complete inner world; an asteroid nothing more than a honey-comb fortress sliding through the void unchallenged. Perfect front for a raider who had little equipment and was short-handed of men. Gunther simply waited for specific cargo ships to rocket by, pulled them or knocked them down and swarmed over them for cargo. The animals served simply to insure against suspicion and the swarms of tourists that filled the void these days. Small fry weren't wanted. They were scared off. The telepathic sending station for the animals was a great bank of intricate, glittering machine, through which strips of colored film with images slid into slots and machine mouths that translated them into thought-emanations. A damned neat piece of genius. "So here we are, still not much better off than we were," growled Irish. "We haven't a ship or a space-radio, and more guards'll turn up any moment. You think we could refocus this doohingey, project the monsters inside the asteroid to fool the pirates themselves?" "What good would that do?" Hathaway gnawed his lip. "They wouldn't fool the engineers who created them, you nut." Marnagan exhaled disgustedly. "Ah, if only the U.S. Cavalry would come riding over the hill—" "Irish!" Hathaway snapped that, his face lighting up. "Irish. The U.S. Cavalry it is!" His eyes darted over the machines. "Here. Help me. We'll stage everything on the most colossal raid of the century." Marnagan winced. "You breathing oxygen or whiskey?" "There's only one stipulation I make, Irish. I want a complete picture of Marnagan capturing Raider's Base. I want a picture of Gunther's face when you do it. Snap it, now, we've got rush work to do. How good an actor are you?" "That's a silly question." "You only have to do three things. Walk with your gun out in front of you, firing. That's number one. Number two is to clutch at your heart and fall down dead. Number three is to clutch at your side, fall down and twitch on the ground. Is that clear?" "Clear as the Coal Sack Nebula...." An hour later Hathaway trudged down a passageway that led out into a sort of city street inside the asteroid. There were about six streets, lined with cube houses in yellow metal, ending near Hathaway in a wide, green-lawned Plaza. Hathaway, weaponless, idly carrying his camera in one hand, walked across the Plaza as if he owned it. He was heading for a building that was pretentious enough to be Gunther's quarters. He got halfway there when he felt a gun in his back. He didn't resist. They took him straight ahead to his destination and pushed him into a room where Gunther sat. Hathaway looked at him. "So you're Gunther?" he said, calmly. The pirate was incredibly old, his bulging forehead stood out over sunken, questioningly dark eyes, and his scrawny body was lost in folds of metal-link cloth. He glanced up from a paper-file, surprised. Before he could speak, Hathaway said: "Everything's over with, Mr. Gunther. The Patrol is in the city now and we're capturing your Base. Don't try to fight. We've a thousand men against your eighty-five." Gunther sat there, blinking at Hathaway, not moving. His thin hands twitched in his lap. "You are bluffing," he said, finally, with a firm directness. "A ship hasn't landed here for an hour. Your ship was the last. Two people were on it. The last I saw of them they were being pursued to the death by the Beasts. One of you escaped, it seemed." "Both. The other guy went after the Patrol." "Impossible!" "I can't respect your opinion, Mr. Gunther." A shouting rose from the Plaza. About fifty of Gunther's men, lounging on carved benches during their time-off, stirred to their feet and started yelling. Gunther turned slowly to the huge window in one side of his office. He stared, hard. The Patrol was coming! Across the Plaza, marching quietly and decisively, came the Patrol. Five hundred Patrolmen in one long, incredible line, carrying paralysis guns with them in their tight hands. Gunther babbled like a child, his voice a shrill dagger in the air. "Get out there, you men! Throw them back! We're outnumbered!" Guns flared. But the Patrol came on. Gunther's men didn't run, Hathaway had to credit them on that. They took it, standing. Hathaway chuckled inside, deep. What a sweet, sweet shot this was. His camera whirred, clicked and whirred again. Nobody stopped him from filming it. Everything was too wild, hot and angry. Gunther was throwing a fit, still seated at his desk, unable to move because of his fragile, bony legs and their atrophied state. Some of the Patrol were killed. Hathaway chuckled again as he saw three of the Patrolmen clutch at their hearts, crumple, lie on the ground and twitch. God, what photography! Gunther raged, and swept a small pistol from his linked corselet. He fired wildly until Hathaway hit him over the head with a paper-weight. Then Hathaway took a picture of Gunther slumped at his desk, the chaos taking place immediately outside his window. The pirates broke and fled, those that were left. A mere handful. And out of the chaos came Marnagan's voice, "Here!" | B. Human imagination |
What effect do the bombs have on the war?
A. They end the war but turn the world into a zombie landscape.
B. They end he war and restore peace and harmony, even though there are still some stragglers wandering home from the war.
C. They end the war, but turn it into a semi-apocalyptic landscape.
D. They end the war, but turn the world into tribal groups with strict borders.
| HOMECOMING BY MIGUEL HIDALGO What lasts forever? Does love? Does death?... Nothing lasts forever.... Not even forever [Transcriber's Note: This etext was produced from Worlds of If Science Fiction, April 1958. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed.] The large horse plodded slowly over the shifting sand. The rider was of medium size, with huge, strong hands and seemingly hollow eyes. Strange eyes, alive and aflame. They had no place in the dust-caked, tired body, yet there they were, seeking, always seeking—searching the clear horizon, and never seeming to find what they sought. The horse moved faster now. They were nearing a river; the water would be welcome on tired bodies and dry throats. He spurred his horse, and when they reached the water's edge, he dismounted and unsaddled the horse. Then both man and horse plunged headlong into the waiting torrent, deep into the cool embrace of the clear liquid. They soaked it into their pores and drank deeply of it, feeling life going once more through their veins. Satisfied, they lifted themselves from the water, and the man lay down on the yellow sand of the river bank to sleep. When he awoke, the sun was almost setting. The bright shafts of red light spilled across the sky, making the mountains silent scarlet shadows on the face of the rippling water. Quickly he gathered driftwood, and built a small fire. From his pack he removed some of the coffee he had found in one of the ruined cities. He brought water from the river in the battered coffee-pot he had salvaged, and while he waited for it to boil, he went to his horse, Conqueror, stroking his mane and whispering in his ear. Then he led him silently to a grassy slope where he hobbled him and left him for the night. In the fading light, he ate the hard beef jerky and drank the scalding coffee. Refreshed and momentarily content, he sat staring into the dying fire, seeing the bright glowing coals as living fingers clutching at the wood in consuming embrace, taking all and returning nothing but ashes. Slowly his eyelids yielded. His body sagged, and blood seemed to fill his brain, bathing it in a gentle, warm flood. He slept. His brain slept. But the portion of his brain called memory stirred. It was all alone; all else was at rest. Images began to appear, drawn from inexhaustible files, wherein are kept all thoughts, past, present, and future.... It was the night before he was to go overseas. World War III had been declared, and he had enlisted, receiving his old rank of captain. He was with his wife in the living room of their home. They had put the children to bed—their sons—and now sat on the couch, watching the blazing fire. It was then that he had showed it to her. "I've got something to tell you, and something to show you." He had removed the box from his pocket and opened it. And heard her cry of surprised joy. "Oh, a ring, and it's a diamond, too!" she cried in her rich, happy voice which always seemed to send a thrill through his body. "It's for you; so long as you wear it, I'll come back, even from the dead, if need be. Read the inscription." She held the ring up to the light and read aloud, "It is forever." Then she had slipped the ring on her finger and her arms around him. He held her very close, feeling the warmth from her body flowing into his and making him oblivious to everything except that she was there in his arms and that he was sinking deep, deep into a familiar sea, where he had been many times before but each time found something new and unexplored, some vastly different emotion he could never quite explain. "Wait!" she cried. "I've something for you, too." She took off the locket she wore about her neck and held it up to the shimmering light, letting it spin at the end of its chain. It caught the shadows of the fire and reflected them, greatly magnified, over the room. It was in the shape of a star, encrusted with emeralds, with one large ruby in the center. When he opened it, he found a picture of her in one side, and in the other a picture of the children. He took her in his arms again, and loosened her long, black hair, burying his face in it for a moment. Then he kissed her, and instantly was drawn down into the abyss which seemed to have no beginning or any end. The next morning had been bleak and gray. The mist clung to the wet, sodden ground, and the air was heavy in his lungs. He had driven off in the jeep the army had sent for him, watching her there on the porch until the mist swirled around her feet and she ran back into the house and slammed the door. His cold fingers found the locket, making a little bulge under his uniform, and the touch of it seemed to warm the blood in his veins. Three days later they had landed in Spain, merged with another division, then crossed the Pyrenees into France, and finally to Paris where the fighting had begun. Already the city was a silent graveyard, littered with the rubble of towers and cathedrals which had once been great. Three years later they were on the road to Moscow. Over a thousand miles lay behind, a dead man on every foot of those miles. Yet victory was near. The Russians had not yet used the H-bomb; the threat of annihilation by the retaliation forces had been too great. He had done well in the war, and had been decorated many times for bravery in action. Now he felt the victory that seemed to be in the air, and he had wished it would come quickly, so that he might return to her. Home. The very feel of the word was everything a battle-weary soldier needed to make him fight harder and live longer. Suddenly he had become aware of a droning, wooshing sound above him. It grew louder and louder until he knew what it was. "Heavy bombers!" The alarm had sounded, and the men had headed for their foxholes. But the planes had passed over, the sun glinting on their bellies, reflecting a blinding light. They were bound for bigger, more important targets. When the all-clear had sounded, the men clambered from their shelters. An icy wind swept the field, bringing with it clouds which covered the sun. A strange fear had gripped him then.... Across the Atlantic, over the pole, via Alaska, the great bombers flew. In cities, great and small, the air raid sirens sounded, high screaming noises which had jarred the people from sleep in time to die. The defending planes roared into the sky to intercept the on-rushing bombers. The horrendous battle split the universe. Many bombers fell, victims of fanatical suicide planes, or of missiles that streaked across the sky which none could escape. But too many bombers got through, dropping their deadly cargo upon the helpless cities. And not all the prayers or entreaties to any God had stopped their carnage. First there had been the red flashes that melted buildings into molten streams, and then the great triple-mushroom cloud filled with the poisonous gases that the wind swept away to other cities, where men had not died quickly and mercifully, but had rotted away, leaving shreds of putrid flesh behind to mark the places where they had crawled. The retaliatory forces had roared away to bomb the Russian cities. Few, if any, had returned. Too much blood and life were on their hands. Those who had remained alive had found a resting place on the crown of some distant mountain. Others had preferred the silent peaceful sea, where flesh stayed not long on bones, and only darting fishes and merciful beams of filtered light found their aluminum coffins. The war had ended. To no avail. Neither side had won. Most of the cities and the majority of the population of both countries had been destroyed. Even their governments had vanished, leaving a silent nothingness. The armies that remained were without leaders, without sources of supplies, save what they could forage and beg from an unfriendly people. They were alone now, a group of tired, battered men, for whom life held nothing. Their families had long since died, their bodies turned to dust, their spirits fled on the winds to a new world. Yet these remnants of an army must return—or at least try. Their exodus was just beginning. Somehow he had managed to hold together the few men left from his force. He had always nourished the hope that she might still be alive. And now that the war was over he had to return—had to know whether she was still waiting for him. They had started the long trek. Throughout Europe anarchy reigned. He and his men were alone. All they could do now was fight. Finally they reached the seaport city of Calais. With what few men he had left, he had commandeered a small yacht, and they had taken to the sea. After months of storms and bad luck, they had been shipwrecked somewhere off the coast of Mexico. He had managed to swim ashore, and had been found by a fisherman's family. Many months he had spent swimming and fishing, recovering his strength, inquiring about the United States. The Mexicans had spoken with fear of the land across the Rio Grande. All its great cities had been destroyed, and those that had been only partially destroyed were devoid of people. The land across the Rio Grande had become a land of shadows. The winds were poisoned, and the few people who might have survived, were crazed and maimed by the blasts. Few men had dared cross the Rio Grande into "El Mundo gris de Noviembre"—the November world. Those who had, had never returned. In time he had traveled north until he reached the Rio Grande. He had waded into the muddy waters and somehow landed on the American side. In the November world. It was rightly called. The deserts were long. All plant life had died, leaving to those once great fertile stretches, nothing but the sad, temporal beauty that comes with death. No people had he seen. Only the ruins of what had once been their cities. He had walked through them, and all that he had seen were the small mutant rodents, and all that he had heard was the occasional swish of the wind as it whisked along what might have been dead leaves, but wasn't. He had been on the trail for a long time. His food was nearly exhausted. The mountains were just beginning, and he hoped to find food there. He had not found food, but his luck had been with him. He had found a horse. Not a normal horse, but a mutation. It was almost twice as large as a regular horse. Its skin seemed to shimmer and was like glassy steel to the touch. From the center of its forehead grew a horn, straight out, as the horn of a unicorn. But most startling of all were the animal's eyes which seemed to speak—a silent mental speech, which he could understand. The horse had looked up as he approached it and seemed to say: "Follow me." And he had followed. Over a mountain, until they came to a pass, and finally to a narrow path which led to an old cabin. He had found it empty, but there were cans of food and a rifle and many shells. He had remained there a long time—how long he could not tell, for he could only measure time by the cycles of the sun and the moon. Finally he had taken the horse, the rifle and what food was left, and once again started the long journey home. The farther north he went, the more life seemed to have survived. He had seen great herds of horses like his own, stampeding across the plains, and strange birds which he could not identify. Yet he had seen no human beings. But he knew he was closer now. Closer to home. He recognized the land. How, he did not know, for it was much changed. A sensing, perhaps, of what it had once been. He could not be more than two days' ride away. Once he was through this desert, he would find her, he would be with her once again; all would be well, and his long journey would be over. The images faded. Even memory slept in a flow of warm blood. Body and mind slept into the shadows of the dawn. He awoke and stretched the cramped muscles of his body. At the edge of the water he removed his clothes and stared at himself in the rippling mirror. His muscles were lean and hard, evenly placed throughout the length of his frame. A deep ridge ran down the length of his torso, separating the muscles, making the chest broad. Well satisfied with his body, he plunged into the cold water, deep down, until he thought his lungs would burst; then swiftly returned to the clean air, tingling in every pore. He dried himself and dressed. Conqueror was eating the long grass near the stream. Quickly he saddled him. No time for breakfast. He would ride all day and the next night. And he would be home. Still northward. The hours crawled slower than a dying man. The sun was a torch that pierced his skin, seeming to melt his bones into a burning stream within his body. But day at last gave way to night, and the sun to the moon. The torch became a white pock-marked goddess, with streaming hair called stars. In the moonlight he had not seen the crater until he was at its very edge. Even then he might not have seen it had not the horse stopped suddenly. The wind swirled through its vast emptiness, slapping his face with dusty hands. For a moment he thought he heard voices—mournful, murmuring voices, echoing up from the misty depths. He turned quickly away and did not look back. Night paled into day; day burned into night. There were clouds in the sky now, and a gentle wind caressed the sweat from his tired body. He stopped. There it was! Barely discernible through the moonlight, he saw it. Home. Quickly he dismounted and ran. Now he could see a small light in the window, and he knew they were there. His breath came in hard ragged gulps. At the window he peered in, and as his eyes became accustomed to the inner gloom, he saw how bare the room was. No matter. Now that he was home he would build new furniture, and the house would be even better than it had been before. Then he saw her. She was sitting motionless in a straight wooden chair beside the fireplace, the feeble light cast by the embers veiling her in mauve shadows. He waited, wondering if she were.... Presently she stirred like a restless child in sleep, then moved from the chair to the pile of wood near the hearth, and replenished the fire. The wood caught quickly, sending up long tongues of flame, and forming a bright pool of light around her. His blood froze. The creature illuminated by the firelight was a monster. Large greasy scales covered its face and arms, and there was no hair on its head. Its gums were toothless cavities in a sunken, mumbling mouth. The eyes, turned momentarily toward the window, were empty of life. "No, no!" he cried soundlessly. This was not his house. In his delirium he had only imagined he had found it. He had been searching so long. He would go on searching. He was turning wearily away from the window when the movement of the creature beside the fire held his attention. It had taken a ring from one skeleton-like finger and stood, turning the ring slowly as if trying to decipher some inscription inside it. He knew then. He had come home. Slowly he moved toward the door. A great weakness was upon him. His feet were stones, reluctant to leave the earth. His body was a weed, shriveled by thirst. He grasped the doorknob and clung to it, looking up at the night sky and trying to draw strength from the wind that passed over him. It was no use. There was no strength. Only fear—a kind of fear he had never known. He fumbled at his throat, his fingers crawling like cold worms around his neck until he found the locket and the clasp which had held it safely through endless nightmare days and nights. He slipped the clasp and the locket fell into his waiting hand. As one in a dream, he opened it, and stared at the pictures, now in the dim moonlight no longer faces of those he loved, but grey ghosts from the past. Even the ruby had lost its glow. What had once been living fire was now a dull glob of darkness. "Nothing is forever!" He thought he had shouted the words, but only a thin sound, the sound of leaves ruffled by the wind, came back to him. He closed the locket and fastened the clasp, and hung it on the doorknob. It moved slowly in the wind, back and forth, like a pendulum. "Forever—forever. Only death is forever." He could have sworn he heard the words. He ran. Away from the house. To the large horse with a horn in the center of its forehead, like a unicorn. Once in the saddle, the spurt of strength left him. His shoulders slumped, his head dropped onto his chest. Conqueror trotted away, the sound of his hooves echoing hollowly in the vast emptiness. | C. They end the war, but turn it into a semi-apocalyptic landscape. |
Generally, which of the following best describes Crystal's character?
A. Kind, quiet, and persistent
B. Naive, fun, and brave
C. Focused, bold, and charismatic
D. Focused, meek, and understanding
| MONOPOLY By Vic Phillips and Scott Roberts Sheer efficiency and good management can make a monopoly grow into being. And once it grows, someone with a tyrant mind is going to try to use it as a weapon if he can— [Transcriber's Note: This etext was produced from Astounding Science-Fiction April 1942. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed.] "That all, chief? Gonna quit now?" Brian Hanson looked disgustedly at Pete Brent, his lanky assistant. That was the first sign of animation he had displayed all day. "I am, but you're not," Hanson told him grimly. "Get your notes straightened up. Run those centrifuge tests and set up the still so we can get at that vitamin count early in the morning." "Tomorrow morning? Aw, for gosh sakes, chief, why don't you take a day off sometime, or better yet, a night off. It'd do you good to relax. Boy, I know a swell blonde you could go for. Wait a minute, I've got her radiophone number somewhere—just ask for Myrtle." Hanson shrugged himself out of his smock. "Never mind Myrtle, just have that equipment set up for the morning. Good night." He strode out of the huge laboratory, but his mind was still on the vitamin research they had been conducting, he barely heard the remarks that followed him. "One of these days the chief is going to have his glands catch up with him." "Not a chance," Pete Brent grunted. Brian Hanson wondered dispassionately for a moment how his assistants could fail to be as absorbed as he was by the work they were doing, then he let it go as he stepped outside the research building. He paused and let his eyes lift to the buildings that surrounded the compound. This was the administrative heart of Venus City. Out here, alone, he let his only known emotion sweep through him, pride. He had an important role in the building of this great new city. As head of the Venus Consolidated Research Organization, he was in large part responsible for the prosperity of this vigorous, young world. Venus Consolidated had built up this city and practically everything else that amounted to anything on this planet. True, there had been others, pioneers, before the company came, who objected to the expansion of the monopolistic control. But, if they could not realize that the company's regime served the best interests of the planet, they would just have to suffer the consequences of their own ignorance. There had been rumors of revolution among the disgruntled older families. He heard there had been killings, but that was nonsense. Venus Consolidated police had only powers of arrest. Anything involving executions had to be referred to the Interplanetary Council on Earth. He dismissed the whole business as he did everything else that did not directly influence his own department. He ignored the surface transport system and walked to his own apartment. This walk was part of a regular routine of physical exercise that kept his body hard and resilient in spite of long hours spent in the laboratory. As he opened the door of his apartment he heard the water running into his bath. Perfect timing. He was making that walk in precisely seven minutes, four and four-fifths seconds. He undressed and climbed into the tub, relaxing luxuriously in the exhilaration of irradiated water. He let all the problems of his work drift away, his mind was a peaceful blank. Then someone was hammering on his head. He struggled reluctantly awake. It was the door that was being attacked, not his head. The battering thunder continued persistently. He swore and sat up. "What do you want?" There was no answer; the hammering continued. "All right! All right! I'm coming!" He yelled, crawled out of the tub and reached for his bathrobe. It wasn't there. He swore some more and grabbed a towel, wrapping it inadequately around him; it didn't quite meet astern. He paddled wetly across the floor sounding like a flock of ducks on parade. Retaining the towel with one hand he inched the door cautiously open. "What the devil—" He stopped abruptly at the sight of a policeman's uniform. "Sorry, sir, but one of those rebels is loose in the Administration Center somewhere. We're making a check-up of all the apartments." "Well, you can check out; I haven't got any blasted rebels in here." The policeman's face hardened, then relaxed knowingly. "Oh, I see, sir. No rebels, of course. Sorry to have disturbed you. Have a good—Good night, sir," he saluted and left. Brian closed the door in puzzlement. What the devil had that flat-foot been smirking about? Well, maybe he could get his bath now. Hanson turned away from the door and froze in amazement. Through the open door of his bedroom he could see his bed neatly turned down as it should be, but the outline under the counterpane and the luxuriant mass of platinum-blond hair on the pillow was certainly no part of his regular routine. "Hello." The voice matched the calm alertness of a pair of deep-blue eyes. Brian just stared at her in numbed fascination. That was what the policeman had meant with his insinuating smirk. "Just ask for Myrtle." Pete Brent's joking words flashed back to him. Now he got it. This was probably the young fool's idea of a joke. He'd soon fix that. "All right, joke's over, you can beat it now." "Joke? I don't see anything funny, unless it's you and that suggestive towel. You should either abandon it or get one that goes all the way round." Brian slowly acquired a complexion suitable for painting fire plugs. "Shut up and throw me my dressing gown." He gritted. The girl swung her legs out of bed and Brian blinked; she was fully dressed. The snug, zippered overall suit she wore did nothing to conceal the fact that she was a female. He wrapped his bathrobe austerely around him. "Well, now what?" she asked and looked at him questioningly. "Well, what do you think?" he burst out angrily. "I'm going to finish my bath and I'd suggest you go down to the laboratory and hold hands with Pete. He'd appreciate it." He got the impression that the girl was struggling heroically to refrain from laughing and that didn't help his dignity any. He strode into the bathroom, slammed the door and climbed back into the bath. The door opened a little. "Well, good-by now." The girl said sweetly. "Remember me to the police force." "Get out of here!" he yelled and the door shut abruptly on a rippling burst of laughter. Damn women! It was getting so a man had to pack a gun with him or something. And Pete Brent. He thought with grim satisfaction of the unending extra work that was going to occur around the laboratory from now on. He sank back into the soothing liquid embrace of the bath and deliberately set his mind loose to wander in complete relaxation. A hammering thunder burst on the outer door. He sat up with a groan. "Lay off, you crazy apes!" he yelled furiously, but the pounding continued steadily. He struggled out of the bath, wrapped his damp bathrobe clammily around him and marched to the door with a seething fury of righteous anger burning within him. He flung the door wide, his mouth all set for a withering barrage, but he didn't get a chance. Four police constables and a sergeant swarmed into the room, shoving him away from the door. "Say! What the—" "Where is she?" the sergeant demanded. "Wherethehell's who?" "Quit stallin', bud. You know who. That female rebel who was in here." "Rebel? You're crazy! That was just ... Pete said ... rebel? Did you say rebel?" "Yeah, I said rebel, an' where is she?" "She ... why ... why ... she left, of course. You don't think I was going to have women running around in here, do you?" "She wuz in his bed when I seen her, sarge," one of the guards contributed. "But she ain't there now." "You don't think that I—" "Listen, bud, we don't do the thinkin' around here. You come on along and see the chief." Brian had had about enough. "I'm not going anywhere to see anybody. Maybe you don't know who I am. You can't arrest me." Brian Hanson, Chief of Research for Venus Consolidated, as dignified as possible in a damp bathrobe, glared out through the bars at a slightly bewildered Pete Brent. "What the devil do you want? Haven't you caused enough blasted trouble already?" "Me? For gosh sakes, chief—" "Yes, you! If sending that damn blonde to my apartment and getting me arrested is your idea of a joke—" "But, my gosh, I didn't send anybody, chief. And this is no joke. That wasn't Myrtle, that was Crystal James, old man James' daughter. They're about the oldest family on Venus. Police have been after her for months; she's a rebel and she's sure been raising plenty of hell around here. She got in and blew out the main communications control panel last night. Communications been tied up all day." Pete lowered his voice to an appreciative whisper, "Gosh, chief, I didn't know you had it in you. How long have you been in with that bunch? Is that girl as good-looking as they say she is?" "Now listen here, Brent. I don't know—" "Oh, it's all right, chief. You can trust me. I won't give you away." "There's nothing to give away, you fool!" Brian bellowed. "I don't know anything about any damn rebels. All I want is to get out of here—" "Gotcha, chief," Brent whispered understandingly. "I'll see if I can pass the word along." "Come here, you idiot!" Brian screamed after his erstwhile assistant. "Pipe down there, bud," a guard's voice cut in chillingly. Brian retired to his cell bunk and clutched his aching head in frustrated fury. For the nineteenth time Brian Hanson strode to the door of his cell and rattled the bars. "Listen here, guard, you've got to take a message to McHague. You can't hold me here indefinitely." "Shut up. Nobody ain't takin' no message to McHague. I don't care if you are—" Brian's eyes almost popped out as he saw a gloved hand reach around the guard's neck and jam a rag over his nose and mouth. Swift shadows moved expertly before his astonished gaze. Another guard was caught and silenced as he came around the end of the corridor. Someone was outside his cell door, a hooded figure which seemed, somehow, familiar. "Hello, pantless!" a voice breathed. He knew that voice! "What the devil are you doing here?" "Somebody by the name of Pete Brent tipped us off that you were in trouble because of me. But don't worry, we're going to get you out." "Damn that fool kid! Leave me alone. I don't want to get out of here that way!" he yelled wildly. "Guards! Help!" "Shut up! Do you want to get us shot?" "Sure I do. Guards! Guards!" Someone came running. "Guards are coming," a voice warned. He could hear the girl struggling with the lock. "Damn," she swore viciously. "This is the wrong key! Your goose is sure cooked now. Whether you like it or not, you'll hang with us when they find us trying to get you out of here." Brian felt as though something had kicked him in the stomach. She was right! He had to get out now. He wouldn't be able to explain this away. "Give me that key," he hissed and grabbed for it. He snapped two of the coigns off in the lock and went to work with the rest of the key. He had designed these escape-proof locks himself. In a few seconds the door swung open and they were fleeing silently down the jail corridor. The girl paused doubtfully at a crossing passage. "This way," he snarled and took the lead. He knew the ground plan of this jail perfectly. He had a moment of wonder at the crazy spectacle of himself, the fair-haired boy of Venus Consolidated, in his flapping bathrobe, leading a band of escaping rebels out of the company's best jail. They burst around a corner onto a startled guard. "They're just ahead of us," Brian yelled. "Come on!" "Right with you," the guard snapped and ran a few steps with them before a blackjack caught up with him and he folded into a corner. "Down this way, it's a short cut." Brian led the way to a heavily barred side door. The electric eye tripped a screaming alarm, but the broken key in Brian's hands opened the complicated lock in a matter of seconds. They were outside the jail on a side street, the door closed and the lock jammed immovably behind them. Sirens wailed. The alarm was out! The street suddenly burst into brilliance as the floodlights snapped on. Brian faltered to a stop and Crystal James pushed past him. "We've got reinforcements down here," she said, then skidded to a halt. Two guards barred the street ahead of them. Brian felt as though his stomach had fallen down around his ankles and was tying his feet up. He couldn't move. The door was jammed shut behind them, they'd have to surrender and there'd be no explaining this break. He started mentally cursing Pete Brent, when a projector beam slashed viciously by him. These guards weren't fooling! He heard a gasping grunt of pain as one of the rebels went down. They were shooting to kill. He saw a sudden, convulsive movement from the girl. A black object curved out against the lights. The sharp, ripping blast of an atomite bomb thundered along the street and slammed them to the ground. The glare left them blinded. He struggled to his feet. The guards had vanished, a shallow crater yawned in the road where they had been. "We've got to run!" the girl shouted. He started after her. Two surface transport vehicles waited around the corner. Brian and the rebels bundled into them and took away with a roar. The chase wasn't organized yet, and they soon lost themselves in the orderly rush of Venus City traffic. The two carloads of rebels cruised nonchalantly past the Administration Center and pulled into a private garage a little beyond. "What are we stopping here for?" Brian demanded. "We've got to get away." "That's just what we're doing," Crystal snapped. "Everybody out." The rebels piled out and the cars pulled away to become innocuous parts of the traffic stream. The rebels seemed to know where they were going and that gave them the edge on Brian. They followed Crystal down into the garage's repair pit. She fumbled in the darkness a moment, then a darker patch showed as a door swung open in the side of the pit. They filed into the solid blackness after her and the door thudded shut. The beam of a torch stabbed through the darkness and they clambered precariously down a steep, steel stairway. "Where the dickens are we?" Brian whispered hoarsely. "Oh, you don't have to whisper, we're safe enough here. This is one of the air shafts leading down to the old mines." "Old mines? What old mines?" "That's something you newcomers don't know anything about. This whole area was worked out long before Venus Consolidated came to the planet. These old tunnels run all under the city." They went five hundred feet down the air shaft before they reached a level tunnel. "What do we do? Hide here?" "I should say not. Serono Zeburzac, head of McHague's secret police will be after us now. We won't be safe anywhere near Venus City." "Don't be crazy. That Serono Zeburzac stuff is just a legend McHague keeps up to scare people with." "That's what you think," Crystal snapped. "McHague's legend got my father and he'll get all of us unless we run the whole company right off the planet." "Well, what the dickens does he look like?" Brian asked doubtfully. "I don't know, but his left hand is missing. Dad did some good shooting before he died," she said grimly. Brian was startled at the icy hardness of her voice. Two of the rebels pulled a screening tarpaulin aside and revealed one of the old-type ore cars that must have been used in the ancient mines. A brand-new atomic motor gleamed incongruously at one end. The rebels crowded into it and they went rumbling swiftly down the echoing passage. The lights of the car showed the old working, rotten and crumbling, fallen in in some places and signs of new work where the rebels had cleared away the debris of years. Brian struggled into a zippered overall suit as they followed a twisting, tortuous course for half an hour, switching from one tunnel to another repeatedly until he had lost all conception of direction. Crystal James, at the controls, seemed to know exactly where they were going. The tunnel emerged in a huge cavern that gloomed darkly away in every direction. The towering, massive remains of old machinery, eroded and rotten with age crouched like ancient, watching skeletons. "These were the old stamp mills," the girl said, and her voice seemed to be swallowed to a whisper in the vast, echoing darkness. Between two rows of sentinel ruins they came suddenly on two slim Venusian atmospheric ships. Dim light spilled over them from a ragged gash in the wall of the cavern. Brian followed Crystal into the smaller of the two ships and the rest of the rebels manned the other. "Wait a minute, how do we get out of here?" Brian demanded. "Through that hole up there," the girl said matter-of-factly. "You're crazy, you can't get through there." "Oh, yeah? Just watch this." The ship thundered to life beneath them and leaped off in a full-throttled take-off. "We're going to crash! That gap isn't wide enough!" The sides of the gap rushed in on the tips of the stubby wings. Brian braced himself for the crash, but it didn't come. At the last possible second, the ship rolled smoothly over. At the moment it flashed through the opening it was stood vertically on edge. Crystal held the ship in its roll and completed the maneuver outside the mountain while Brian struggled to get his internal economy back into some semblance of order. "That's some flying," he said as soon as he could speak. Crystal looked at him in surprise. "That's nothing. We Venusians fly almost as soon as we can walk." "Oh—I see," Brian said weakly and a few moments later he really did see. Two big, fast, green ships, carrying the insignia of the Venus Consolidated police, cruised suddenly out from a mountain air station. An aërial torpedo exploded in front of the rebel ship. Crystal's face set in grim lines as she pulled the ship up in a screaming climb. Brian got up off the floor. "You don't have to get excited like that," he complained. "They weren't trying to hit us." "That's what you think," Crystal muttered. "Those children don't play for peanuts." "But, girl, they're just Venus Consolidated police. They haven't got any authority to shoot anyone." "Authority doesn't make much difference to them," Crystal snapped bitterly. "They've been killing people all over the planet. What do you think this revolution is about?" "You must be mistak—" He slumped to the floor as Crystal threw the ship into a mad, rolling spin. A tremendous crash thundered close astern. "I guess that was a mistake!" Crystal yelled as she fought the controls. Brian almost got to his feet when another wild maneuver hurled him back to the floor. The police ship was right on their tail. The girl gunned her craft into a snap Immelmann and swept back on their pursuers, slicing in close over the ship. Brian's eyes bulged as he saw a long streak of paint and metal ripped off the wing of the police ship. He saw the crew battling their controls in startled terror. The ship slipped frantically away and fell into a spin. "That's them," Crystal said with satisfaction. "How are the others doing?" "Look! They're hit!" Brian felt sick. The slower rebel freight ship staggered drunkenly as a torpedo caught it and ripped away half a wing. It plunged down in flames with the white flowers of half a dozen parachutes blossoming around it. Brian watched in horror as the police ship came deliberately about. They heard its forward guns go into action. The bodies of the parachutists jerked and jumped like crazy marionettes as the bullets smashed into them. It was over in a few moments. The dead rebels drifted down into the mist-shrouded depths of the valley. "The dirty, murdering rats!" Brian's voice ripped out in a fury of outrage. "They didn't have a chance!" "Don't get excited," Crystal told him in a dead, flat voice. "That's just normal practice. If you'd stuck your nose out of your laboratory once in a while, you'd have heard of these things." "But why—" He ducked away instinctively as a flight of bullets spanged through the fuselage. "They're after us now!" Crystal's answer was to yank the ship into a rocketing climb. The police were watching for that. The big ship roared up after them. "Just follow along, suckers," Crystal invited grimly. She snapped the ship into a whip stall. For one nauseating moment they hung on nothing, then the ship fell over on its back and they screamed down in a terminal velocity dive, heading for the safety of the lower valley mists. The heavier police ship, with its higher wing-loading, could not match the maneuver. The rebel craft plunged down through the blinding fog. Half-seen, ghostly fingers of stone clutched up at them, talons of gray rock missed and fell away again as Crystal nursed the ship out of its dive. " Phew! " Brian gasped. "Well, we got away that time. How in thunder can you do it?" "Well, you don't do it on faith. Take a look at that fuel gauge! We may get as far as our headquarters—or we may not." For twenty long minutes they groped blindly through the fog, flying solely by instruments and dead reckoning. The needle of the fuel gauge flickered closer and closer to the danger point. They tore loose from the clinging fog as it swung firmly to "Empty." The drive sputtered and coughed and died. "That's figuring it nice and close," Crystal said in satisfaction. "We can glide in from here." "Into where?" Brian demanded. All he could see immediately ahead was the huge bulk of a mountain which blocked the entire width of the valley and soared sheer up to the high-cloud level. His eyes followed it up and up— "Look! Police ships. They've seen us!" "Maybe they haven't. Anyway, there's only one place we can land." The ship lunged straight for the mountain wall! "Are you crazy? Watch out—we'll crash!" "You leave the flying to me," Crystal snapped. She held the ship in its glide, aiming directly for the tangled foliage of the mountain face. Brian yelped and cowered instinctively back. The lush green of the mountainside swirled up to meet them. They ripped through the foliage—there was no crash. They burst through into a huge, brilliantly lighted cavern and settled to a perfect landing. Men came running. Crystal tumbled out of her ship. "Douse those lights," she shouted. "The police are outside." A tall, lean man with bulbous eyes and a face like a startled horse, rushed up to Crystal. "What do you mean by leading them here?" he yelled, waving his hands. "They jumped us when we had no fuel, and quit acting like an idiot." The man was shaking, his eyes looked wild. "They'll kill us. We've got to get out of here." "Wait, you fool. They may not even have seen us." But he was gone, running toward a group of ships lined up at the end of the cavern. "Who was that crazy coot and what is this place?" Brian demanded. "That was Gort Sterling, our leader," the girl said bitterly. "And this is our headquarters." One of the ships at the back of the cavern thundered to life, streaked across the floor and burst out through the opening Crystal's ship had left. "He hasn't got a chance! We'll be spotted for sure, now." The other rebels waited uncertainly, but not for long. There was the crescendoing roar of ships in a dive followed by the terrific crash of an explosion. "They got him!" Crystal's voice was a moan. "Oh, the fool, the fool!" "Sounded like more than one ship. They'll be after us, now. Is there any other way of getting out of this place?" "Not for ships. We'll have to walk and they'll follow us." "We've got to slow them down some way, then. I wonder how the devil they traced us? I thought we lost them in that fog." "It's that Serono Zeburzac, the traitor. He knows these mountains as well as we do." "How come?" "The Zeburzacs are one of the old families, but he sold out to McHague." "Well, what do we do now? Just stand here? It looks like everybody's leaving." "We might as well just wait," Crystal said hopelessly. "It won't do us any good to run out into the hills. Zeburzac and his men will follow." "We could slow them down some by swinging a couple of those ships around so their rocket exhausts sweep the entrance to the cavern," Brian suggested doubtfully. She looked at him steadily. "You sound like the only good rebel left. We can try it, anyway." They ran two ships out into the middle of the cavern, gunned them around and jockeyed them into position—not a moment too soon. Half a dozen police showed in brief silhouette as they slipped cautiously into the cavern, guns ready, expecting resistance. They met a dead silence. A score or more followed them without any attempt at concealment. Then Brian and Crystal cut loose with the drives of the two ships. Startled screams of agony burst from the crowded group of police as they were caught in the annihilating cross fire of roaring flame. They crisped and twisted, cooked to scorched horrors before they fell. A burst of thick, greasy smoke rushed out of the cavern. Two of the police, their clothes and flesh scorched and flaming, plunged as shrieking, living torches down the mountainside. Crystal was white and shaking, her face set in a mask of horror, as she climbed blindly from her ship. "Let's get away! I can smell them burning," she shuddered and covered her face with her hands. Brian grabbed her and shook her. "Snap out of it," he barked. "That's no worse than shooting helpless men in parachutes. We can't go, yet; we're not finished here." "Oh, let them shoot us! I can't go through that again!" "You don't have to. Wait here." He climbed back into one of the ships and cut the richness of the fuel mixture down till the exhaust was a lambent, shuddering stutter, verging on extinction. He dashed to the other ship and repeated the maneuver, fussing with the throttle till he had the fuel mixture adjusted to critical fineness. The beat of the stuttering exhaust seemed to catch up to the other and built to an aching pulsation. In a moment the whole mass of air in the cavern hit the frequency with a subtle, intangible thunder of vibration. Crystal screamed. "Brian! There's more police cutting in around the entrance." Brian clambered out of the ship and glanced at the glowing points in the rock where the police were cutting their way through outside the line of the exhaust flames. The pulsating thunder in the cavern crescendoed to an intolerable pitch. A huge mass of stalactites crashed to the floor. "It's time to check out," Brian shouted. Crystal led the way as they fled down the escape tunnel. The roaring crash of falling rock was a continuous, increasing avalanche of sound in the cavern behind them. They emerged from the tunnel on the face of the mountain, several hundred yards to the east of the cavern entrance. The ground shook and heaved beneath them. "The whole side of the mountain's sliding," Crystal screamed. "Run!" Brian shoved her and they plunged madly through the thick tangle of jungle away from the slide. Huge boulders leaped and smashed through the matted bush around them. Crystal went down as the ground slipped from under her. Brian grabbed her and a tree at the same time. The tree leaned and crashed down the slope, the whole jungle muttered and groaned and came to life as it joined the roaring rush of the slide. They were tumbled irresistibly downward, riding the edge of the slide for terrifying minutes till it stilled and left them bruised and shaken in a tangle of torn vegetation. The remains of two police ships, caught without warning in the rush as they attempted to land, stuck up grotesquely out of the foot of the slide. The dust was settling away. A flock of brilliant blue, gliding lizards barking in raucous terror, fled down the valley. Then they were gone and the primeval silence settled back into place. Brian and Crystal struggled painfully to solid ground. Crystal gazed with a feeling of awe at the devastated mountainside. "How did you do it?" "It's a matter of harmonics," Brian explained. "If you hit the right vibratory combination, you can shake anything down. But now that we've made a mess of the old homestead, what do we do?" "Walk," Crystal said laconically. She led the way as they started scrambling through the jungle up the mountainside. "Where are we heading for?" Brian grunted as he struggled along. "The headquarters of the Carlton family. They're the closest people we can depend on. They've kept out of the rebellion, but they're on our side. They've helped us before." | C. Focused, bold, and charismatic |
What is Lockheed Martin's 2 year total revenue CAGR from FY2020 to FY2022 (in units of percents and round to one decimal place)? Provide a response to the question by primarily using the statement of income. | Evidence 0:
Lockheed Martin Corporation
Consolidated Statements of Earnings
(in millions, except per share data)
Years Ended December 31,
2022
2021
2020
Net sales
Products
$
55,466 $
56,435 $
54,928
Services
10,518
10,609
10,470
Total net sales
65,984
67,044
65,398
Cost of sales
Products
(49,577)
(50,273)
(48,996)
Services
(9,280)
(9,463)
(9,371)
Severance and other charges
(100)
(36)
(27)
Other unallocated, net
1,260
1,789
1,650
Total cost of sales
(57,697)
(57,983)
(56,744)
Gross profit
8,287
9,061
8,654
Other income (expense), net
61
62
(10)
Operating profit
8,348
9,123
8,644
Interest expense
(623)
(569)
(591)
Non-service FAS pension (expense) income
(971)
(1,292)
219
Other non-operating (expense) income, net
(74)
288
(37)
Earnings from continuing operations before income taxes
6,680
7,550
8,235
Income tax expense
(948)
(1,235)
(1,347)
Net earnings from continuing operations
5,732
6,315
6,888
Net loss from discontinued operations
(55)
Net earnings
$
5,732 $
6,315 $
6,833
Earnings (loss) per common share
Basic
Continuing operations
$
21.74 $
22.85 $
24.60
Discontinued operations
(0.20)
Basic earnings per common share
$
21.74 $
22.85 $
24.40
Diluted
Continuing operations
$
21.66 $
22.76 $
24.50
Discontinued operations
(0.20)
Diluted earnings per common share
$
21.66 $
22.76 $
24.30
The accompanying notes are an integral part of these consolidated financial statements.
Table of Contents
63 | 0.4% |
What gave Starre the right to claim the asteroid as her own when Parker and Queazy arrived?
A. She had made a deal with the Interplanetary Commission.
B. Her grandfather had purchased the asteroid for her.
C. By common law, asteroids up to a certain size belong to whoever happens to be on them.
D. She had signed an interplanetary lease agreement.
| COSMIC YO-YO By ROSS ROCKLYNNE "Want an asteroid in your backyard? We supply cheap. Trouble also handled without charge." Interplanetary Hauling Company. (ADVT.) [Transcriber's Note: This etext was produced from Planet Stories Summer 1945. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed.] Bob Parker, looking through the photo-amplifiers at the wedge-shaped asteroid, was plainly flabbergasted. Not in his wildest imaginings had he thought they would actually find what they were looking for. "Cut the drive!" he yelled at Queazy. "I've got it, right on the nose. Queazy, my boy, can you imagine it? We're in the dough. Not only that, we're rich! Come here!" Queazy discharged their tremendous inertia into the motive-tubes in such a manner that the big, powerful ship was moving at the same rate as the asteroid below—47.05 miles per second. He came slogging back excitedly, put his eyes to the eyepiece. He gasped, and his big body shook with joyful ejaculations. "She checks down to the last dimension," Bob chortled, working with slide-rule and logarithm tables. "Now all we have to do is find out if she's made of tungsten, iron, quartz crystals, and cinnabar! But there couldn't be two asteroids of that shape anywhere else in the Belt, so this has to be it!" He jerked a badly crumpled ethergram from his pocket, smoothed it out, and thumbed his nose at the signature. "Whee! Mr. Andrew S. Burnside, you owe us five hundred and fifty thousand dollars!" Queazy straightened. A slow, likeable smile wreathed his tanned face. "Better take it easy," he advised, "until I land the ship and we use the atomic whirl spectroscope to determine the composition of the asteroid." "Have it your way," Bob Parker sang, happily. He threw the ethergram to the winds and it fell gently to the deck-plates. While Queazy—so called because his full name was Quentin Zuyler—dropped the ship straight down to the smooth surface of the asteroid, and clamped it tight with magnetic grapples, Bob flung open the lazarette, brought out two space-suits. Moments later, they were outside the ship, with star-powdered infinity spread to all sides. In the ship, the ethergram from Andrew S. Burnside, of Philadelphia, one of the richest men in the world, still lay on the deck-plates. It was addressed to: Mr. Robert Parker, President Interplanetary Hauling & Moving Co., 777 Main Street, Satterfield City, Fontanaland, Mars. The ethergram read: Received your advertising literature a week ago. Would like to state that yes I would like an asteroid in my back yard. Must meet following specifications: 506 feet length, long enough for wedding procession; 98 feet at base, tapering to 10 feet at apex; 9-12 feet thick; topside smooth-plane, underside rough-plane; composed of iron ore, tungsten, quartz crystals, and cinnabar. Must be in my back yard before 11:30 A.M. my time, for important wedding June 2, else order is void. Will pay $5.00 per ton. Bob Parker had received that ethergram three weeks ago. And if The Interplanetary Hauling & Moving Co., hadn't been about to go on the rocks (chiefly due to the activities of Saylor & Saylor, a rival firm) neither Bob nor Queazy would have thought of sending an answering ethergram to Burnside stating that they would fill the order. It was, plainly, a hair-brained request. And yet, if by some chance there was such a rigidly specified asteroid, their financial worries would be over. That they had actually discovered the asteroid, using their mass-detectors in a weight-elimination process, seemed like an incredible stroke of luck. For there are literally millions of asteroids in the asteroid belt, and they had been out in space only three weeks. The "asteroid in your back yard" idea had been Bob Parker's originally. Now it was a fad that was sweeping Earth, and Burnside wasn't the first rich man who had decided to hold a wedding on top of an asteroid. Unfortunately, other interplanetary moving companies had cashed in on that brainstorm, chiefly the firm of the Saylor brothers—which persons Bob Parker intended to punch in the nose some day. And would have before this if he hadn't been lanky and tall while they were giants. Now that he and Queazy had found the asteroid, they were desperate to get it to its destination, for fear that the Saylor brothers might get wind of what was going on, and try to beat them out of their profits. Which was not so far-fetched, because the firm of Saylor & Saylor made no pretense of being scrupulous. Now they scuffed along the smooth-plane topside of the asteroid, the magnets in their shoes keeping them from stepping off into space. They came to the broad base of the asteroid-wedge, walked over the edge and "down" the twelve-foot thickness. Here they squatted, and Bob Parker happily clamped the atomic-whirl spectroscope to the rough surface. By the naked eye, they could see iron ore, quartz crystals, cinnabar, but he had the spectroscope and there was no reason why he shouldn't use it. He satisfied himself as to the exterior of the asteroid, and then sent the twin beams deep into its heart. The beams crossed, tore atoms from molecules, revolved them like an infinitely fine powder. The radiations from the sundered molecules traveled back up the beams to the atomic-whirl spectroscope. Bob watched a pointer which moved slowly up and up—past tungsten, past iridium, past gold— Bob Parker said, in astonishment, "Hell! There's something screwy about this business. Look at that point—" Neither he nor Queazy had the opportunity to observe the pointer any further. A cold, completely disagreeable feminine voice said, "May I ask what you interlopers are doing on my asteroid?" Bob started so badly that the spectroscope's settings were jarred and the lights in its interior died. Bob twisted his head around as far as he could inside the "aquarium"—the glass helmet, and found himself looking at a space-suited girl who was standing on the edge of the asteroid "below." "Ma'am," said Bob, blinking, "did you say something?" Queazy made a gulping sound and slowly straightened. He automatically reached up as if he would take off his hat and twist it in his hands. "I said," remarked the girl, "that you should scram off of my asteroid. And quit poking around at it with that spectroscope. I've already taken a reading. Cinnabar, iron ore, quartz crystals, tungsten. Goodbye." Bob's nose twitched as he adjusted his glasses, which he wore even inside his suit. He couldn't think of anything pertinent to say. He knew that he was slowly working up a blush. Mildly speaking, the girl was beautiful, and though only her carefully made-up face was visible—cool blue eyes, masterfully coiffed, upswept, glinting brown hair, wilful lips and chin—Bob suspected the rest of her compared nicely. Her expression darkened as she saw the completely instinctive way he was looking at her and her radioed-voice rapped out, "Now you two boys go and play somewhere else! Else I'll let the Interplanetary Commission know you've infringed the law. G'bye!" She turned and disappeared. Bob awoke from his trance, shouted desperately, "Hey! Wait! You! " He and Queazy caught up with her on the side of the asteroid they hadn't yet examined. It was a rough plane, completing the rigid qualifications Burnside had set down. "Wait a minute," Bob Parker begged nervously. "I want to make some conversation, lady. I'm sure you don't understand the conditions—" The girl turned and drew a gun from a holster. It was a spasticizer, and it was three times as big as her gloved hand. "I understand conditions better than you do," she said. "You want to move this asteroid from its orbit and haul it back to Earth. Unfortunately, this is my home, by common law. Come back in a month. I don't expect to be here then." "A month!" Parker burst the word out. He started to sweat, then his face became grim. He took two slow steps toward the girl. She blinked and lost her composure and unconsciously backed up two steps. About twenty steps away was her small dumbbell-shaped ship, so shiny and unscarred that it reflected starlight in highlights from its curved surface. A rich girl's ship, Bob Parker thought angrily. A month would be too late! He said grimly, "Don't worry. I don't intend to pull any rough stuff. I just want you to listen to reason. You've taken a whim to stay on an asteroid that doesn't mean anything to you one way or another. But to us—to me and Queazy here—it means our business. We got an order for this asteroid. Some screwball millionaire wants it for a backyard wedding see? We get five hundred and fifty thousand dollars for it! If we don't take this asteroid to Earth before June 2, we go back to Satterfield City and work the rest of our lives in the glass factories. Don't we, Queazy?" Queazy said simply, "That's right, miss. We're in a spot. I assure you we didn't expect to find someone living here." The girl holstered her spasticizer, but her completely inhospitable expression did not change. She put her hands on the bulging hips of her space-suit. "Okay," she said. "Now I understand the conditions. Now we both understand each other. G'bye again. I'm staying here and—" she smiled sweetly "—it may interest you to know that if I let you have the asteroid you'll save your business, but I'll meet a fate worse than death! So that's that." Bob recognized finality when he saw it. "Come on, Queazy," he said fuming. "Let this brat have her way. But if I ever run across her without a space-suit on I'm going to give her the licking of her life, right where it'll do the most good!" He turned angrily, but Queazy grabbed his arm, his mouth falling open. He pointed off into space, beyond the girl. "What's that?" he whispered. "What's wha— Oh! " Bob Parker's stomach caved in. A few hundred feet away, floating gently toward the asteroid, came another ship—a ship a trifle bigger than their own. The girl turned, too. They heard her gasp. In another second, Bob was standing next to her. He turned the audio-switch to his headset off, and spoke to the girl by putting his helmet against hers. "Listen to me, miss," he snapped earnestly, when she tried to draw away. "Don't talk by radio. That ship belongs to the Saylor brothers! Oh, Lord, that this should happen! Somewhere along the line, we've been double-crossed. Those boys are after this asteroid too, and they won't hesitate to pull any rough stuff. We're in this together, understand? We got to back each other up." The girl nodded dumbly. Suddenly she seemed to be frightened. "It's—it's very important that this—this asteroid stay right where it is," she said huskily. "What—what will they do?" Bob Parker didn't answer. The big ship had landed, and little blue sparks crackled between the hull and the asteroid as the magnetic clamps took hold. A few seconds later, the airlocks swung down, and five men let themselves down to the asteroid's surface and stood surveying the three who faced them. The two men in the lead stood with their hands on their hips; their darkish, twin faces were grinning broadly. "A pleasure," drawled Wally Saylor, looking at the girl. "What do you think of this situation Billy?" "It's obvious," drawled Billy Saylor, rocking back and forth on his heels, "that Bob Parker and company have double-crossed us. We'll have to take steps." The three men behind the Saylor twins broke into rough, chuckling laughter. Bob Parker's gorge rose. "Scram," he said coldly. "We've got an ethergram direct from Andrew S. Burnside ordering this asteroid." "So have we," Wally Saylor smiled—and his smile remained fixed, dangerous. He started moving forward, and the three men in back came abreast, forming a semi-circle which slowly closed in. Bob Parker gave back a step, as he saw their intentions. "We got here first," he snapped harshly. "Try any funny stuff and we'll report you to the Interplanetary Commission!" It was Bob Parker's misfortune that he didn't carry a weapon. Each of these men carried one or more, plainly visible. But he was thinking of the girl's spasticizer—a paralyzing weapon. He took a hair-brained chance, jerked the spasticizer from the girl's holster and yelled at Queazy. Queazy got the idea, urged his immense body into motion. He hurled straight at Billy Saylor, lifted him straight off the asteroid and threw him away, into space. He yelled with triumph. At the same time, the spasticizer Bob held was shot cleanly out of his hand by Wally Saylor. Bob roared, started toward Wally Saylor, knocked the smoking gun from his hand with a sweeping arm. Then something crushing seemed to hit him in the stomach, grabbing at his solar plexus. He doubled up, gurgling with agony. He fell over on his back, and his boots were wrenched loose from their magnetic grip. Vaguely, before the flickering points of light in his brain subsided to complete darkness, he heard the girl's scream of rage—then a scream of pain. What had happened to Queazy he didn't know. He felt so horribly sick, he didn't care. Then—lights out. Bob Parker came to, the emptiness of remote starlight in his face. He opened his eyes. He was slowly revolving on an axis. Sometimes the Sun swept across his line of vision. A cold hammering began at the base of his skull, a sensation similar to that of being buried alive. There was no asteroid, no girl, no Queazy. He was alone in the vastness of space. Alone in a space-suit. "Queazy!" he whispered. "Queazy! I'm running out of air!" There was no answer from Queazy. With sick eyes, Bob studied the oxygen indicator. There was only five pounds pressure. Five pounds! That meant he had been floating around out here—how long? Days at least—maybe weeks! It was evident that somebody had given him a dose of spastic rays, enough to screw up every muscle in his body to the snapping point, putting him in such a condition of suspended animation that his oxygen needs were small. He closed his eyes, trying to fight against panic. He was glad he couldn't see any part of his body. He was probably scrawny. And he was hungry! "I'll starve," he thought. "Or suffocate to death first!" He couldn't keep himself from taking in great gulps of air. Minutes, then hours passed. He was breathing abnormally, and there wasn't enough air in the first place. He pleaded continually for Queazy, hoping that somehow Queazy could help, when probably Queazy was in the same condition. He ripped out wild curses directed at the Saylor brothers. Murderers, both of them! Up until this time, he had merely thought of them as business rivals. If he ever got out of this— He groaned. He never would get out of it! After another hour, he was gasping weakly, and yellow spots danced in his eyes. He called Queazy's name once more, knowing that was the last time he would have strength to call it. And this time the headset spoke back! Bob Parker made a gurgling sound. A voice came again, washed with static, far away, burbling, but excited. Bob made a rattling sound in his throat. Then his eyes started to close, but he imagined that he saw a ship, shiny and small, driving toward him, growing in size against the backdrop of the Milky Way. He relapsed, a terrific buzzing in his ears. He did not lose consciousness. He heard voices, Queazy's and the girl's, whoever she was. Somebody grabbed hold of his foot. His "aquarium" was unbuckled and good air washed over his streaming face. The sudden rush of oxygen to his brain dizzied him. Then he was lying on a bunk, and gradually the world beyond his sick body focussed in his clearing eyes and he knew he was alive—and going to stay that way, for awhile anyway. "Thanks, Queazy," he said huskily. Queazy was bending over him, his anxiety clearing away from his suddenly brightening face. "Don't thank me," he whispered. "We'd have both been goners if it hadn't been for her. The Saylor brothers left her paralyzed like us, and when she woke up she was on a slow orbit around her ship. She unstrapped her holster and threw it away from her and it gave her enough reaction to reach the ship. She got inside and used the direction-finder on the telaudio and located me first. The Saylors scattered us far and wide." Queazy's broad, normally good-humored face twisted blackly. "The so and so's didn't care if we lived or died." Bob saw the girl now, standing a little behind Queazy, looking down at him curiously, but unhappily. Her space-suit was off. She was wearing lightly striped blue slacks and blue silk blouse and she had a paper flower in her hair. Something in Bob's stomach caved in as his eyes widened on her. The girl said glumly, "I guess you men won't much care for me when you find out who I am and what I've done. I'm Starre Lowenthal—Andrew S. Burnside's granddaughter!" Bob came slowly to his feet, and matched Queazy's slowly growing anger. "Say that again?" he snapped. "This is some kind of dirty trick you and your grandfather cooked up?" "No!" she exclaimed. "No. My grandfather didn't even know there was an asteroid like this. But I did, long before he ordered it from you—or from the Saylor brothers. You see—well, my granddad's about the stubbornest old hoot-owl in this universe! He's always had his way, and when people stand in his way, that's just a challenge to him. He's been badgering me for years to marry Mac, and so has Mac—" "Who's Mac?" Queazy demanded. "My fiancé, I guess," she said helplessly. "He's one of my granddad's protégés. Granddad's always financing some likely young man and giving him a start in life. Mac has become pretty famous for his Mercurian water-colors—he's an artist. Well, I couldn't hold out any longer. If you knew my grandfather, you'd know how absolutely impossible it is to go against him when he's got his mind set! I was just a mass of nerves. So I decided to trick him and I came out to the asteroid belt and picked out an asteroid that was shaped so a wedding could take place on it. I took the measurements and the composition, then I told my grandfather I'd marry Mac if the wedding was in the back yard on top of an asteroid with those measurements and made of iron ore, tungsten, and so forth. He agreed so fast he scared me, and just to make sure that if somebody did find the asteroid in time they wouldn't be able to get it back to Earth, I came out here and decided to live here. Asteroids up to a certain size belong to whoever happens to be on them, by common law.... So I had everything figured out—except," she added bitterly, "the Saylor brothers! I guess Granddad wanted to make sure the asteroid was delivered, so he gave the order to several companies." Bob swore under his breath. He went reeling across to a port, and was gratified to see his and Queazy's big interplanetary hauler floating only a few hundred feet away. He swung around, looked at Queazy. "How long were we floating around out there?" "Three weeks, according to the chronometer. The Saylor boys gave us a stiff shot." " Ouch! " Bob groaned. Then he looked at Starre Lowenthal with determination. "Miss, pardon me if I say that this deal you and your granddad cooked up is plain screwy! With us on the butt end. But I'm going to put this to you plainly. We can catch up with the Saylor brothers even if they are three weeks ahead of us. The Saylor ship and ours both travel on the HH drive—inertia-less. But the asteroid has plenty of inertia, and so they'll have to haul it down to Earth by a long, spiraling orbit. We can go direct and probably catch up with them a few hundred thousand miles this side of Earth. And we can have a fling at getting the asteroid back!" Her eyes sparkled. "You mean—" she cried. Then her attractive face fell. "Oh," she said. " Oh! And when you get it back, you'll land it." "That's right," Bob said grimly. "We're in business. For us, it's a matter of survival. If the by-product of delivering the asteroid is your marriage—sorry! But until we do get the asteroid back, we three can work as a team if you're willing. We'll fight the other problem out later. Okay?" She smiled tremulously. "Okay, I guess." Queazy looked from one to another of them. He waved his hand scornfully at Bob. "You're plain nuts," he complained. "How do you propose to go about convincing the Saylor brothers they ought to let us have the asteroid back? Remember, commercial ships aren't allowed to carry long-range weapons. And we couldn't ram the Saylor brothers' ship—not without damaging our own ship just as much. Go ahead and answer that." Bob looked at Queazy dismally. "The old balance-wheel," he groaned at Starre. "He's always pulling me up short when I go off half-cocked. All I know is, that maybe we'll get a good idea as we go along. In the meantime, Starre—ahem—none of us has eaten in three weeks...?" Starre got the idea. She smiled dazzlingly and vanished toward the galley. Bob Parker was in love with Starre Lowenthal. He knew that after five days out, as the ship hurled itself at breakneck speed toward Earth; probably that distracting emotion was the real reason he couldn't attach any significance to Starre's dumbbell-shaped ship, which trailed astern, attached by a long cable. Starre apparently knew he was in love with her, too, for on the fifth day Bob was teaching her the mechanics of operating the hauler, and she gently lifted his hand from a finger-switch. "Even I know that isn't the control to the Holloway vacuum-feeder, Bob. That switch is for the—ah—the anathern tube, you told me. Right?" "Right," he said unsteadily. "Anyway, Starre, as I was saying, this ship operates according to the reverse Fitzgerald Contraction Formula. All moving bodies contract in the line of motion. What Holloway and Hammond did was to reverse that universal law. They caused the contraction first—motion had to follow! The gravitonic field affects every atom in the ship with the same speed at the same time. We could go from zero speed to our top speed of two thousand miles a second just like that!" He snapped his fingers. "No acceleration effects. This type of ship, necessary in our business, can stop flat, back up, ease up, move in any direction, and the passengers wouldn't have any feeling of motion at—Oh, hell!" Bob groaned, the serious glory of her eyes making him shake. He took her hand. "Starre," he said desperately, "I've got to tell you something—" She jerked her hand away. "No," she exclaimed in an almost frightened voice. "You can't tell me. There's—there's Mac," she finished, faltering. "The asteroid—" "You have to marry him?" Her eyes filled with tears. "I have to live up to the bargain." "And ruin your whole life," he ground out. Suddenly, he turned back to the control board, quartered the vision plate. He pointed savagely to the lower left quarter, which gave a rearward view of the dumbbell ship trailing astern. "There's your ship, Starre." He jabbed his finger at it. "I've got a feeling—and I can't put the thought into concrete words—that somehow the whole solution of the problem of grabbing the asteroid back lies there. But how? How? " Starre's blue eyes followed the long cable back to where it was attached around her ship's narrow midsection. She shook her head helplessly. "It just looks like a big yo-yo to me." "A yo-yo?" "Yes, a yo-yo. That's all." She was belligerent. "A yo-yo !" Bob Parker yelled the word and almost hit the ceiling, he got out of the chair so fast. "Can you imagine it! A yo-yo!" He disappeared from the room. "Queazy!" he shouted. " Queazy, I've got it! " It was Queazy who got into his space-suit and did the welding job, fastening two huge supra-steel "eyes" onto the dumbbell-shaped ship's narrow midsection. Into these eyes cables which trailed back to two winches in the big ship's nose were inserted, welded fast, and reinforced. The nose of the hauler was blunt, perfectly fitted for the job. Bob Parker practiced and experimented for three hours with this yo-yo of cosmic dimensions, while Starre and Queazy stood over him bursting into strange, delighted squeals of laughter whenever the yo-yo reached the end of its double cable and started rolling back up to the ship. Queazy snapped his fingers. "It'll work!" His gray eyes showed satisfaction. "Now, if only the Saylor brothers are where we calculated!" They weren't where Bob and Queazy had calculated, as they had discovered the next day. They had expected to pick up the asteroid on their mass-detectors a few hundred thousand miles outside of the Moon's orbit. But now they saw the giant ship attached like a leech to the still bigger asteroid—inside the Moon's orbit! A mere two hundred thousand miles from Earth! "We have to work fast," Bob stammered, sweating. He got within naked-eye distance of the Saylor brothers' ship. Below, Earth was spread out, a huge crescent shape, part of the Eastern hemisphere vaguely visible through impeding clouds and atmosphere. The enemy ship was two miles distant, a black shadow occulting part of the brilliant sky. It was moving along a down-spiraling path toward Earth. Queazy's big hand gripped his shoulder. "Go to it, Bob!" Bob nodded grimly. He backed the hauler up about thirty miles, then sent it forward again, directly toward the Saylor brothers' ship at ten miles per second. And resting on the blunt nose of the ship was the "yo-yo." There was little doubt the Saylors' saw their approach. But, scornfully, they made no attempt to evade. There was no possible harm the oncoming ship could wreak. Or at least that was what they thought, for Bob brought the hauler's speed down to zero—and Starre Lowenthal's little ship, possessing its own inertia, kept on moving! It spun away from the hauler's blunt nose, paying out two rigid lengths of cable behind it as it unwound, hurled itself forward like a fantastic spinning cannon ball. "It's going to hit!" The excited cry came from Starre. But Bob swore. The dumbbell ship reached the end of its cables, falling a bare twenty feet short of completing its mission. It didn't stop spinning, but came winding back up the cable, at the same terrific speed with which it had left. Bob sweated, having only fractions of seconds in which to maneuver for the "yo-yo" could strike a fatal blow at the hauler too. It was ticklish work completely to nullify the "yo-yo's" speed. Bob used exactly the same method of catching the "yo-yo" on the blunt nose of the ship as a baseball player uses to catch a hard-driven ball in his glove—namely, by matching the ball's speed and direction almost exactly at the moment of impact. And now Bob's hours of practice paid dividends, for the "yo-yo" came to rest snugly, ready to be released again. All this had happened in such a short space of time that the Saylor brothers must have had only a bare realization of what was going on. But by the time the "yo-yo" was flung at them again, this time with better calculations, they managed to put the firmly held asteroid between them and the deadly missile. But it was clumsy evasion, for the asteroid was several times as massive as the ship which was towing it, and its inertia was great. And as soon as the little ship came spinning back to rest, Bob flung the hauler to a new vantage point and again the "yo-yo" snapped out. And this time—collision! Bob yelled as he saw the stern section of the Saylor brothers' ship crumple like tissue paper crushed between the hand. The dumbbell-shaped ship, smaller, and therefore stauncher due to the principle of the arch, wound up again, wobbling a little. It had received a mere dent in its starboard half. Starre was chortling with glee. Queazy whispered, "Attaboy, Bob! This time we'll knock 'em out of the sky!" The "yo-yo" came to rest and at the same moment a gong rang excitedly. Bob knew what that meant. The Saylor brothers were trying to establish communication. Queazy was across the room in two running strides. He threw in the telaudio and almost immediately, Wally Saylor's big body built up in the plate. Wally Saylor's face was quivering with wrath. "What do you damned fools think you're trying to do?" he roared. "You've crushed in our stern section. You've sliced away half of our stern jets. Air is rushing out! You'll kill us!" "Now," Bob drawled, "you're getting the idea." "I'll inform the Interplanetary Commission!" screamed Saylor. " If you're alive," Bob snarled wrathfully. "And you won't be unless you release the asteroid." "I'll see you in Hades first!" "Hades," remarked Bob coldly, "here you come!" He snapped the hauler into its mile-a-second speed again, stopped it at zero. And the "yo-yo" went on its lone, destructive sortie. For a fraction of a second Wally Saylor exhibited the countenance of a doomed man. In the telaudio plate, he whirled, and diminished in size with a strangled yell. The "yo-yo" struck again, but Bob Parker maneuvered its speed in such a manner that it struck in the same place as before, but not as heavily, then rebounded and came spinning back with perfect, sparkling precision. And even before it snugged itself into its berth, it was apparent that the Saylor brothers had given up. Like a wounded terrier, their ship shook itself free of the asteroid, hung in black space for a second, then vanished with a flaming puff of released gravitons from its still-intact jets. The battle was won! | C. By common law, asteroids up to a certain size belong to whoever happens to be on them. |
All of the following motivate Prantera to accept the proposal from Brett-James and Reston-Ferrell EXCEPT:
A. He does not need to worry about Temple-Tracy's followers seeking revenge
B. He does not have to fear being arrested by the police
C. He is unlikely to encounter someone with weapons during the job
D. He does not have a chance of being sent back to 1960
| Illustrated by van Dongen A gun is an interesting weapon; it can be hired, of course, and naturally doesn't care who hires it. Something much the same can be said of the gunman, too.... GUN FOR HIRE By MACK REYNOLDS Joe Prantera called softly, "Al." The pleasurable, comfortable, warm feeling began spreading over him, the way it always did. The older man stopped and squinted, but not suspiciously, even now. The evening was dark, it was unlikely that the other even saw the circle of steel that was the mouth of the shotgun barrel, now resting on the car's window ledge. "Who's it?" he growled. Joe Prantera said softly, "Big Louis sent me, Al." And he pressed the trigger. And at that moment, the universe caved inward upon Joseph Marie Prantera. There was nausea and nausea upon nausea. There was a falling through all space and through all time. There was doubling and twisting and twitching of every muscle and nerve. There was pain, horror and tumultuous fear. And he came out of it as quickly and completely as he'd gone in. He was in, he thought, a hospital and his first reaction was to think, This here California. Everything different. Then his second thought was Something went wrong. Big Louis, he ain't going to like this. He brought his thinking to the present. So far as he could remember, he hadn't completely pulled the trigger. That at least meant that whatever the rap was it wouldn't be too tough. With luck, the syndicate would get him off with a couple of years at Quentin. A door slid open in the wall in a way that Joe had never seen a door operate before. This here California. The clothes on the newcomer were wrong, too. For the first time, Joe Prantera began to sense an alienness—a something that was awfully wrong. The other spoke precisely and slowly, the way a highly educated man speaks a language which he reads and writes fluently but has little occasion to practice vocally. "You have recovered?" Joe Prantera looked at the other expressionlessly. Maybe the old duck was one of these foreign doctors, like. The newcomer said, "You have undoubtedly been through a most harrowing experience. If you have any untoward symptoms, possibly I could be of assistance." Joe couldn't figure out how he stood. For one thing, there should have been some kind of police guard. The other said, "Perhaps a bit of stimulant?" Joe said flatly, "I wanta lawyer." The newcomer frowned at him. "A lawyer?" "I'm not sayin' nothin'. Not until I get a mouthpiece." The newcomer started off on another tack. "My name is Lawrence Reston-Farrell. If I am not mistaken, you are Joseph Salviati-Prantera." Salviati happened to be Joe's mother's maiden name. But it was unlikely this character could have known that. Joe had been born in Naples and his mother had died in childbirth. His father hadn't brought him to the States until the age of five and by that time he had a stepmother. "I wanta mouthpiece," Joe said flatly, "or let me outta here." Lawrence Reston-Farrell said, "You are not being constrained. There are clothes for you in the closet there." Joe gingerly tried swinging his feet to the floor and sitting up, while the other stood watching him, strangely. He came to his feet. With the exception of a faint nausea, which brought back memories of that extreme condition he'd suffered during ... during what? He hadn't the vaguest idea of what had happened. He was dressed in a hospital-type nightgown. He looked down at it and snorted and made his way over to the closet. It opened on his approach, the door sliding back into the wall in much the same manner as the room's door had opened for Reston-Farrell. Joe Prantera scowled and said, "These ain't my clothes." "No, I am afraid not." "You think I'd be seen dead wearing this stuff? What is this, some religious crackpot hospital?" Reston-Farrell said, "I am afraid, Mr. Salviati-Prantera, that these are the only garments available. I suggest you look out the window there." Joe gave him a long, chill look and then stepped to the window. He couldn't figure the other. Unless he was a fruitcake. Maybe he was in some kind of pressure cooker and this was one of the fruitcakes. He looked out, however, not on the lawns and walks of a sanitarium but upon a wide boulevard of what was obviously a populous city. And for a moment again, Joe Prantera felt the depths of nausea. This was not his world. He stared for a long, long moment. The cars didn't even have wheels, he noted dully. He turned slowly and faced the older man. Reston-Farrell said compassionately, "Try this, it's excellent cognac." Joe Prantera stared at him, said finally, flatly, "What's it all about?" The other put down the unaccepted glass. "We were afraid first realization would be a shock to you," he said. "My colleague is in the adjoining room. We will be glad to explain to you if you will join us there." "I wanta get out of here," Joe said. "Where would you go?" The fear of police, of Al Rossi's vengeance, of the measures that might be taken by Big Louis on his failure, were now far away. Reston-Farrell had approached the door by which he had entered and it reopened for him. He went through it without looking back. There was nothing else to do. Joe dressed, then followed him. In the adjoining room was a circular table that would have accommodated a dozen persons. Two were seated there now, papers, books and soiled coffee cups before them. There had evidently been a long wait. Reston-Farrell, the one Joe had already met, was tall and drawn of face and with a chainsmoker's nervousness. The other was heavier and more at ease. They were both, Joe estimated, somewhere in their middle fifties. They both looked like docs. He wondered, all over again, if this was some kind of pressure cooker. But that didn't explain the view from the window. Reston-Farrell said, "May I present my colleague, Citizen Warren Brett-James? Warren, this is our guest from ... from yesteryear, Mr. Joseph Salviati-Prantera." Brett-James nodded to him, friendly, so far as Joe could see. He said gently, "I think it would be Mr. Joseph Prantera, wouldn't it? The maternal linage was almost universally ignored." His voice too gave the impression he was speaking a language not usually on his tongue. Joe took an empty chair, hardly bothering to note its alien qualities. His body seemed to fit into the piece of furniture, as though it had been molded to his order. Joe said, "I think maybe I'll take that there drink, Doc." Reston-Farrell said, "Of course," and then something else Joe didn't get. Whatever the something else was, a slot opened in the middle of the table and a glass, so clear of texture as to be all but invisible, was elevated. It contained possibly three ounces of golden fluid. Joe didn't allow himself to think of its means of delivery. He took up the drink and bolted it. He put the glass down and said carefully, "What's it all about, huh?" Warren Brett-James said soothingly, "Prepare yourself for somewhat of a shock, Mr. Prantera. You are no longer in Los Angeles—" "Ya think I'm stupid? I can see that." "I was about to say, Los Angeles of 1960. Mr. Prantera, we welcome you to Nuevo Los Angeles." "Ta where?" "To Nuevo Los Angeles and to the year—" Brett-James looked at his companion. "What is the date, Old Calendar?" "2133," Reston-Farrell said. "2133 A.D. they would say." Joe Prantera looked from one of them to the other, scowling. "What are you guys talking about?" Warren Brett-James said softly, "Mr. Prantera, you are no longer in the year 1960, you are now in the year 2133." He said, uncomprehendingly, "You mean I been, like, unconscious for—" He let the sentence fall away as he realized the impossibility. Brett-James said gently, "Hardly for one hundred and seventy years, Mr. Prantera." Reston-Farrell said, "I am afraid we are confusing you. Briefly, we have transported you, I suppose one might say, from your own era to ours." Joe Prantera had never been exposed to the concept of time travel. He had simply never associated with anyone who had ever even remotely considered such an idea. Now he said, "You mean, like, I been asleep all that time?" "Not exactly," Brett-James said, frowning. Reston-Farrell said, "Suffice to say, you are now one hundred and seventy-three years after the last memory you have." Joe Prantera's mind suddenly reverted to those last memories and his eyes narrowed dangerously. He felt suddenly at bay. He said, "Maybe you guys better let me in on what's this all about." Reston-Farrell said, "Mr. Prantera, we have brought you from your era to perform a task for us." Joe stared at him, and then at the other. He couldn't believe he was getting through to them. Or, at least, that they were to him. Finally he said, "If I get this, you want me to do a job for you." "That is correct." Joe said, "You guys know the kind of jobs I do?" "That is correct." "Like hell you do. You think I'm stupid? I never even seen you before." Joe Prantera came abruptly to his feet. "I'm gettin' outta here." For the second time, Reston-Farrell said, "Where would you go, Mr. Prantera?" Joe glared at him. Then sat down again, as abruptly as he'd arisen. "Let's start all over again. I got this straight, you brought me, some screwy way, all the way ... here. O.K., I'll buy that. I seen what it looks like out that window—" The real comprehension was seeping through to him even as he talked. "Everybody I know, Jessie, Tony, the Kid, Big Louis, everybody, they're dead. Even Big Louis." "Yes," Brett-James said, his voice soft. "They are all dead, Mr. Prantera. Their children are all dead, and their grandchildren." The two men of the future said nothing more for long minutes while Joe Prantera's mind whirled its confusion. Finally he said, "What's this bit about you wanting me to give it to some guy." "That is why we brought you here, Mr. Prantera. You were ... you are, a professional assassin." "Hey, wait a minute, now." Reston-Farrell went on, ignoring the interruption. "There is small point in denying your calling. Pray remember that at the point when we ... transported you, you were about to dispose of a contemporary named Alphonso Annunziata-Rossi. A citizen, I might say, whose demise would probably have caused small dismay to society." They had him pegged all right. Joe said, "But why me? Why don't you get some heavy from now? Somebody knows the ropes these days." Brett-James said, "Mr. Prantera, there are no professional assassins in this age, nor have there been for over a century and a half." "Well, then do it yourself." Joe Prantera's irritation over this whole complicated mess was growing. And already he was beginning to long for the things he knew—for Jessie and Tony and the others, for his favorite bar, for the lasagne down at Papa Giovanni's. Right now he could have welcomed a calling down at the hands of Big Louis. Reston-Farrell had come to his feet and walked to one of the large room's windows. He looked out, as though unseeing. Then, his back turned, he said, "We have tried, but it is simply not in us, Mr. Prantera." "You mean you're yella?" "No, if by that you mean afraid. It is simply not within us to take the life of a fellow creature—not to speak of a fellow man." Joe snapped: "Everything you guys say sounds crazy. Let's start all over again." Brett-James said, "Let me do it, Lawrence." He turned his eyes to Joe. "Mr. Prantera, in your own era, did you ever consider the future?" Joe looked at him blankly. "In your day you were confronted with national and international, problems. Just as we are today and just as nations were a century or a millennium ago." "Sure, O.K., so we had problems. I know whatcha mean—like wars, and depressions and dictators and like that." "Yes, like that," Brett-James nodded. The heavy-set man paused a moment. "Yes, like that," he repeated. "That we confront you now indicates that the problems of your day were solved. Hadn't they been, the world most surely would have destroyed itself. Wars? Our pedagogues are hard put to convince their students that such ever existed. More than a century and a half ago our society eliminated the reasons for international conflict. For that matter," he added musingly, "we eliminated most international boundaries. Depressions? Shortly after your own period, man awoke to the fact that he had achieved to the point where it was possible to produce an abundance for all with a minimum of toil. Overnight, for all practical purposes, the whole world was industrialized, automated. The second industrial revolution was accompanied by revolutionary changes in almost every field, certainly in every science. Dictators? Your ancestors found, Mr. Prantera, that it is difficult for a man to be free so long as others are still enslaved. Today the democratic ethic has reached a pinnacle never dreamed of in your own era." "O.K., O.K.," Joe Prantera growled. "So everybody's got it made. What I wanta know is what's all this about me giving it ta somebody? If everything's so great, how come you want me to knock this guy off?" Reston-Farrell bent forward and thumped his right index finger twice on the table. "The bacterium of hate—a new strain—has found the human race unprotected from its disease. We had thought our vaccines immunized us." "What's that suppose to mean?" Brett-James took up the ball again. "Mr. Prantera, have you ever heard of Ghengis Khan, of Tamerlane, Alexander, Caesar?" Joe Prantera scowled at him emptily. "Or, more likely, of Napoleon, Hitler, Stalin?" "Sure I heard of Hitler and Stalin," Joe growled. "I ain't stupid." The other nodded. "Such men are unique. They have a drive ... a drive to power which exceeds by far the ambitions of the average man. They are genii in their way, Mr. Prantera, genii of evil. Such a genius of evil has appeared on the current scene." "Now we're getting somewheres," Joe snorted. "So you got a guy what's a little ambitious, like, eh? And you guys ain't got the guts to give it to him. O.K. What's in it for me?" The two of them frowned, exchanged glances. Reston-Farrell said, "You know, that is one aspect we had not considered." Brett-James said to Joe Prantera, "Had we not, ah, taken you at the time we did, do you realize what would have happened?" "Sure," Joe grunted. "I woulda let old Al Rossi have it right in the guts, five times. Then I woulda took the plane back to Chi." Brett-James was shaking his head. "No. You see, by coincidence, a police squad car was coming down the street just at that moment to arrest Mr. Rossi. You would have been apprehended. As I understand Californian law of the period, your life would have been forfeit, Mr. Prantera." Joe winced. It didn't occur to him to doubt their word. Reston-Farrell said, "As to reward, Mr. Prantera, we have already told you there is ultra-abundance in this age. Once this task has been performed, we will sponsor your entry into present day society. Competent psychiatric therapy will soon remove your present—" "Waita minute, now. You figure on gettin' me candled by some head shrinker, eh? No thanks, Buster. I'm going back to my own—" Brett-James was shaking his head again. "I am afraid there is no return, Mr. Prantera. Time travel works but in one direction, with the flow of the time stream. There can be no return to your own era." Joe Prantera had been rocking with the mental blows he had been assimilating, but this was the final haymaker. He was stuck in this squaresville of a world. Joe Prantera on a job was thorough. Careful, painstaking, competent. He spent the first three days of his life in the year 2133 getting the feel of things. Brett-James and Reston-Farrell had been appointed to work with him. Joe didn't meet any of the others who belonged to the group which had taken the measures to bring him from the past. He didn't want to meet them. The fewer persons involved, the better. He stayed in the apartment of Reston-Farrell. Joe had been right, Reston-Farrell was a medical doctor. Brett-James evidently had something to do with the process that had enabled them to bring Joe from the past. Joe didn't know how they'd done it, and he didn't care. Joe was a realist. He was here. The thing was to adapt. There didn't seem to be any hurry. Once the deal was made, they left it up to him to make the decisions. They drove him around the town, when he wished to check the traffic arteries. They flew him about the whole vicinity. From the air, Southern California looked much the same as it had in his own time. Oceans, mountains, and to a lesser extent, deserts, are fairly permanent even against man's corroding efforts. It was while he was flying with Brett-James on the second day that Joe said, "How about Mexico? Could I make the get to Mexico?" The physicist looked at him questioningly. "Get?" he said. Joe Prantera said impatiently, "The getaway. After I give it to this Howard Temple-Tracy guy, I gotta go on the run, don't I?" "I see." Brett-James cleared his throat. "Mexico is no longer a separate nation, Mr. Prantera. All North America has been united into one unit. Today, there are only eight nations in the world." "Where's the nearest?" "South America." "That's a helluva long way to go on a get." "We hadn't thought of the matter being handled in that manner." Joe eyed him in scorn. "Oh, you didn't, huh? What happens after I give it to this guy? I just sit around and wait for the cops to put the arm on me?" Brett-James grimaced in amusement. "Mr. Prantera, this will probably be difficult for you to comprehend, but there are no police in this era." Joe gaped at him. "No police! What happens if you gotta throw some guy in stir?" "If I understand your idiom correctly, you mean prison. There are no prisons in this era, Mr. Prantera." Joe stared. "No cops, no jails. What stops anybody? What stops anybody from just going into some bank, like, and collecting up all the bread?" Brett-James cleared his throat. "Mr. Prantera, there are no banks." "No banks! You gotta have banks!" "And no money to put in them. We found it a rather antiquated method of distribution well over a century ago." Joe had given up. Now he merely stared. Brett-James said reasonably, "We found we were devoting as much time to financial matters in all their endless ramifications—including bank robberies—as we were to productive efforts. So we turned to more efficient methods of distribution." On the fourth day, Joe said, "O.K., let's get down to facts. Summa the things you guys say don't stick together so good. Now, first place, where's this guy Temple-Tracy you want knocked off?" Reston-Farrell and Brett-James were both present. The three of them sat in the living room of the latter's apartment, sipping a sparkling wine which seemed to be the prevailing beverage of the day. For Joe's taste it was insipid stuff. Happily, rye was available to those who wanted it. Reston-Farrell said, "You mean, where does he reside? Why, here in this city." "Well, that's handy, eh?" Joe scratched himself thoughtfully. "You got somebody can finger him for me?" "Finger him?" "Look, before I can give it to this guy I gotta know some place where he'll be at some time. Get it? Like Al Rossi. My finger, he works in Rossi's house, see? He lets me know every Wednesday night, eight o'clock, Al leaves the house all by hisself. O.K., so I can make plans, like, to give it to him." Joe Prantera wound it up reasonably. "You gotta have a finger." Brett-James said, "Why not just go to Temple-Tracy's apartment and, ah, dispose of him?" "Jest walk in, eh? You think I'm stupid? How do I know how many witnesses hangin' around? How do I know if the guy's carryin' heat?" "Heat?" "A gun, a gun. Ya think I'm stupid? I come to give it to him and he gives it to me instead." Dr. Reston-Farrell said, "Howard Temple-Tracy lives alone. He customarily receives visitors every afternoon, largely potential followers. He is attempting to recruit members to an organization he is forming. It would be quite simple for you to enter his establishment and dispose of him. I assure you, he does not possess weapons." Joe was indignant. "Just like that, eh?" he said sarcastically. "Then what happens? How do I get out of the building? Where's my get car parked? Where do I hide out? Where do I dump the heat?" "Dump the heat?" "Get rid of the gun. You want I should get caught with the gun on me? I'd wind up in the gas chamber so quick—" "See here, Mr. Prantera," Brett-James said softly. "We no longer have capital punishment, you must realize." "O.K. I still don't wanta get caught. What is the rap these days, huh?" Joe scowled. "You said they didn't have no jails any more." "This is difficult for you to understand, I imagine," Reston-Farrell told him, "but, you see, we no longer punish people in this era." That took a long, unbelieving moment to sink in. "You mean, like, no matter what they do? That's crazy. Everybody'd be running around giving it to everybody else." "The motivation for crime has been removed, Mr. Prantera," Reston-Farrell attempted to explain. "A person who commits a violence against another is obviously in need of medical care. And, consequently, receives it." "You mean, like, if I steal a car or something, they just take me to a doctor?" Joe Prantera was unbelieving. "Why would anybody wish to steal a car?" Reston-Farrell said easily. "But if I give it to somebody?" "You will be turned over to a medical institution. Citizen Howard Temple-Tracy is the last man you will ever kill, Mr. Prantera." A chillness was in the belly of Joe Prantera. He said very slowly, very dangerously, "You guys figure on me getting caught, don't you?" "Yes," Brett-James said evenly. "Well then, figure something else. You think I'm stupid?" "Mr. Prantera," Dr. Reston-Farrell said, "there has been as much progress in the field of psychiatry in the past two centuries as there has in any other. Your treatment would be brief and painless, believe me." Joe said coldly, "And what happens to you guys? How do you know I won't rat on you?" Brett-James said gently, "The moment after you have accomplished your mission, we plan to turn ourselves over to the nearest institution to have determined whether or not we also need therapy." "Now I'm beginning to wonder about you guys," Joe said. "Look, all over again, what'd'ya wanta give it to this guy for?" The doctor said, "We explained the other day, Mr. Prantera. Citizen Howard Temple-Tracy is a dangerous, atavistic, evil genius. We are afraid for our institutions if his plans are allowed to mature." "Well if you got things so good, everybody's got it made, like, who'd listen to him?" The doctor nodded at the validity of the question. "Mr. Prantera, Homo sapiens is a unique animal. Physically he matures at approximately the age of thirteen. However, mental maturity and adjustment is often not fully realized until thirty or even more. Indeed, it is sometimes never achieved. Before such maturity is reached, our youth are susceptible to romantic appeal. Nationalism, chauvinism, racism, the supposed glory of the military, all seem romantic to the immature. They rebel at the orderliness of present society. They seek entertainment in excitement. Citizen Temple-Tracy is aware of this and finds his recruits among the young." "O.K., so this guy is dangerous. You want him knocked off before he screws everything up. But the way things are, there's no way of making a get. So you'll have to get some other patsy. Not me." "I am afraid you have no alternative," Brett-James said gently. "Without us, what will you do? Mr. Prantera, you do not even speak the language." "What'd'ya mean? I don't understand summa the big words you eggheads use, but I get by O.K." Brett-James said, "Amer-English is no longer the language spoken by the man in the street, Mr. Prantera. Only students of such subjects any longer speak such tongues as Amer-English, French, Russian or the many others that once confused the race with their limitations as a means of communication." "You mean there's no place in the whole world where they talk American?" Joe demanded, aghast. Dr. Reston-Farrell controlled the car. Joe Prantera sat in the seat next to him and Warren Brett-James sat in the back. Joe had, tucked in his belt, a .45 caliber automatic, once displayed in a museum. It had been more easily procured than the ammunition to fit it, but that problem too had been solved. The others were nervous, obviously repelled by the very conception of what they had planned. Inwardly, Joe was amused. Now that they had got in the clutch, the others were on the verge of chickening out. He knew it wouldn't have taken much for them to cancel the project. It wasn't any answer though. If they allowed him to call it off today, they'd talk themselves into it again before the week was through. Besides, already Joe was beginning to feel the comfortable, pleasurable, warm feeling that came to him on occasions like this. He said, "You're sure this guy talks American, eh?" Warren Brett-James said, "Quite sure. He is a student of history." "And he won't think it's funny I talk American to him, eh?" "He'll undoubtedly be intrigued." They pulled up before a large apartment building that overlooked the area once known as Wilmington. Joe was coolly efficient now. He pulled out the automatic, held it down below his knees and threw a shell into the barrel. He eased the hammer down, thumbed on the safety, stuck the weapon back in his belt and beneath the jacketlike garment he wore. He said, "O.K. See you guys later." He left them and entered the building. An elevator—he still wasn't used to their speed in this era—whooshed him to the penthouse duplex occupied by Citizen Howard Temple-Tracy. There were two persons in the reception room but they left on Joe's arrival, without bothering to look at him more than glancingly. He spotted the screen immediately and went over and stood before it. The screen lit and revealed a heavy-set, dour of countenance man seated at a desk. He looked into Joe Prantera's face, scowled and said something. Joe said, "Joseph Salviati-Prantera to interview Citizen Howard Temple-Tracy." The other's shaggy eyebrows rose. "Indeed," he said. "In Amer-English?" Joe nodded. "Enter," the other said. A door had slid open on the other side of the room. Joe walked through it and into what was obviously an office. Citizen Temple-Tracy sat at a desk. There was only one other chair in the room. Joe Prantera ignored it and remained standing. Citizen Temple-Tracy said, "What can I do for you?" Joe looked at him for a long, long moment. Then he reached down to his belt and brought forth the .45 automatic. He moistened his lips. Joe said softly, "You know what this here is?" Temple-Tracy stared at the weapon. "It's a handgun, circa, I would say, about 1925 Old Calendar. What in the world are you doing with it?" Joe said, very slowly, "Chief, in the line you're in these days you needa heavy around with wunna these. Otherwise, Chief, you're gunna wind up in some gutter with a lotta holes in you. What I'm doin', I'm askin' for a job. You need a good man knows how to handle wunna these, Chief." Citizen Howard Temple-Tracy eyed him appraisingly. "Perhaps," he said, "you are right at that. In the near future, I may well need an assistant knowledgeable in the field of violence. Tell me more about yourself. You surprise me considerably." "Sure, Chief. It's kinda a long story, though. First off, I better tell you you got some bad enemies, Chief. Two guys special, named Brett-James and Doc Reston-Farrell. I think one of the first jobs I'm gunna hafta do for you, Chief, is to give it to those two." THE END Transcriber's Note: This etext was produced from Analog December 1960. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed. Minor spelling and typographical errors have been corrected without note. | A. He does not need to worry about Temple-Tracy's followers seeking revenge |
What languages are were included in the dataset of hateful content? | ### Introduction
Social media are sometimes used to disseminate hateful messages. In Europe, the current surge in hate speech has been linked to the ongoing refugee crisis. Lawmakers and social media sites are increasingly aware of the problem and are developing approaches to deal with it, for example promising to remove illegal messages within 24 hours after they are reported BIBREF0 . This raises the question of how hate speech can be detected automatically. Such an automatic detection method could be used to scan the large amount of text generated on the internet for hateful content and report it to the relevant authorities. It would also make it easier for researchers to examine the diffusion of hateful content through social media on a large scale. From a natural language processing perspective, hate speech detection can be considered a classification task: given an utterance, determine whether or not it contains hate speech. Training a classifier requires a large amount of data that is unambiguously hate speech. This data is typically obtained by manually annotating a set of texts based on whether a certain element contains hate speech. The reliability of the human annotations is essential, both to ensure that the algorithm can accurately learn the characteristics of hate speech, and as an upper bound on the expected performance BIBREF1 , BIBREF2 . As a preliminary step, six annotators rated 469 tweets. We found that agreement was very low (see Section 3). We then carried out group discussions to find possible reasons. They revealed that there is considerable ambiguity in existing definitions. A given statement may be considered hate speech or not depending on someone's cultural background and personal sensibilities. The wording of the question may also play a role. We decided to investigate the issue of reliability further by conducting a more comprehensive study across a large number of annotators, which we present in this paper. Our contribution in this paper is threefold: ### Hate Speech
For the purpose of building a classifier, warner2012 define hate speech as “abusive speech targeting specific group characteristics, such as ethnic origin, religion, gender, or sexual orientation”. More recent approaches rely on lists of guidelines such as a tweet being hate speech if it “uses a sexist or racial slur” BIBREF2 . These approaches are similar in that they leave plenty of room for personal interpretation, since there may be differences in what is considered offensive. For instance, while the utterance “the refugees will live off our money” is clearly generalising and maybe unfair, it is unclear if this is already hate speech. More precise definitions from law are specific to certain jurisdictions and therefore do not capture all forms of offensive, hateful speech, see e.g. matsuda1993. In practice, social media services are using their own definitions which have been subject to adjustments over the years BIBREF3 . As of June 2016, Twitter bans hateful conduct. With the rise in popularity of social media, the presence of hate speech has grown on the internet. Posting a tweet takes little more than a working internet connection but may be seen by users all over the world. Along with the presence of hate speech, its real-life consequences are also growing. It can be a precursor and incentive for hate crimes, and it can be so severe that it can even be a health issue BIBREF4 . It is also known that hate speech does not only mirror existing opinions in the reader but can also induce new negative feelings towards its targets BIBREF5 . Hate speech has recently gained some interest as a research topic on the one hand – e.g. BIBREF6 , BIBREF4 , BIBREF7 – but also as a problem to deal with in politics such as the No Hate Speech Movement by the Council of Europe. The current refugee crisis has made it evident that governments, organisations and the public share an interest in controlling hate speech in social media. However, there seems to be little consensus on what hate speech actually is. ### Compiling A Hate Speech Corpus
As previously mentioned, there is no German hate speech corpus available for our needs, especially not for the very recent topic of the refugee crisis in Europe. We therefore had to compile our own corpus. We used Twitter as a source as it offers recent comments on current events. In our study we only considered the textual content of tweets that contain certain keywords, ignoring those that contain pictures or links. This section provides a detailed description of the approach we used to select the tweets and subsequently annotate them. To find a large amount of hate speech on the refugee crisis, we used 10 hashtags that can be used in an insulting or offensive way. Using these hashtags we gathered 13 766 tweets in total, roughly dating from February to March 2016. However, these tweets contained a lot of non-textual content which we filtered out automatically by removing tweets consisting solely of links or images. We also only considered original tweets, as retweets or replies to other tweets might only be clearly understandable when reading both tweets together. In addition, we removed duplicates and near-duplicates by discarding tweets that had a normalised Levenshtein edit distance smaller than .85 to an aforementioned tweet. A first inspection of the remaining tweets indicated that not all search terms were equally suited for our needs. The search term #Pack (vermin or lowlife) found a potentially large amount of hate speech not directly linked to the refugee crisis. It was therefore discarded. As a last step, the remaining tweets were manually read to eliminate those which were difficult to understand or incomprehensible. After these filtering steps, our corpus consists of 541 tweets, none of which are duplicates, contain links or pictures, or are retweets or replies. As a first measurement of the frequency of hate speech in our corpus, we personally annotated them based on our previous expertise. The 541 tweets were split into six parts and each part was annotated by two out of six annotators in order to determine if hate speech was present or not. The annotators were rotated so that each pair of annotators only evaluated one part. Additionally the offensiveness of a tweet was rated on a 6-point Likert scale, the same scale used later in the study. Even among researchers familiar with the definitions outlined above, there was still a low level of agreement (Krippendorff's INLINEFORM0 ). This supports our claim that a clearer definition is necessary in order to be able to train a reliable classifier. The low reliability could of course be explained by varying personal attitudes or backgrounds, but clearly needs more consideration. ### Methods
In order to assess the reliability of the hate speech definitions on social media more comprehensively, we developed two online surveys in a between-subjects design. They were completed by 56 participants in total (see Table TABREF7 ). The main goal was to examine the extent to which non-experts agree upon their understanding of hate speech given a diversity of social media content. We used the Twitter definition of hateful conduct in the first survey. This definition was presented at the beginning, and again above every tweet. The second survey did not contain any definition. Participants were randomly assigned one of the two surveys. The surveys consisted of 20 tweets presented in a random order. For each tweet, each participant was asked three questions. Depending on the survey, participants were asked (1) to answer (yes/no) if they considered the tweet hate speech, either based on the definition or based on their personal opinion. Afterwards they were asked (2) to answer (yes/no) if the tweet should be banned from Twitter. Participants were finally asked (3) to answer how offensive they thought the tweet was on a 6-point Likert scale from 1 (Not offensive at all) to 6 (Very offensive). If they answered 4 or higher, the participants had the option to state which particular words they found offensive. After the annotation of the 20 tweets, participants were asked to voluntarily answer an open question regarding the definition of hate speech. In the survey with the definition, they were asked if the definition of Twitter was sufficient. In the survey without the definition, the participants were asked to suggest a definition themselves. Finally, sociodemographic data were collected, including age, gender and more specific information regarding the participant's political orientation, migration background, and personal position regarding the refugee situation in Europe. The surveys were approved by the ethical committee of the Department of Computer Science and Applied Cognitive Science of the Faculty of Engineering at the University of Duisburg-Essen. ### Preliminary Results and Discussion
Since the surveys were completed by 56 participants, they resulted in 1120 annotations. Table TABREF7 shows some summary statistics. To assess whether the definition had any effect, we calculated, for each participant, the percentage of tweets they considered hate speech or suggested to ban and their mean offensiveness rating. This allowed us to compare the two samples for each of the three questions. Preliminary Shapiro-Wilk tests indicated that some of the data were not normally distributed. We therefore used the Wilcoxon-Mann-Whitney (WMW) test to compare the three pairs of series. The results are reported in Table TABREF7 . Participants who were shown the definition were more likely to suggest to ban the tweet. In fact, participants in group one very rarely gave different answers to questions one and two (18 of 500 instances or 3.6%). This suggests that participants in that group aligned their own opinion with the definition. We chose Krippendorff's INLINEFORM0 to assess reliability, a measure from content analysis, where human coders are required to be interchangeable. Therefore, it measures agreement instead of association, which leaves no room for the individual predilections of coders. It can be applied to any number of coders and to interval as well as nominal data. BIBREF8 This allowed us to compare agreement between both groups for all three questions. Figure FIGREF8 visualises the results. Overall, agreement was very low, ranging from INLINEFORM0 to INLINEFORM1 . In contrast, for the purpose of content analysis, Krippendorff recommends a minimum of INLINEFORM2 , or a minimum of INLINEFORM3 for applications where some uncertainty is unproblematic BIBREF8 . Reliability did not consistently increase when participants were shown a definition. To measure the extent to which the annotations using the Twitter definition (question one in group one) were in accordance with participants' opinions (question one in group two), we calculated, for each tweet, the percentage of participants in each group who considered it hate speech, and then calculated Pearson's correlation coefficient. The two series correlate strongly ( INLINEFORM0 ), indicating that they measure the same underlying construct. ### Conclusion and Future Work
This paper describes the creation of our hate speech corpus and offers first insights into the low agreement among users when it comes to identifying hateful messages. Our results imply that hate speech is a vague concept that requires significantly better definitions and guidelines in order to be annotated reliably. Based on the present findings, we are planning to develop a new coding scheme which includes clear-cut criteria that let people distinguish hate speech from other content. Researchers who are building a hate speech detection system might want to collect multiple labels for each tweet and average the results. Of course this approach does not make the original data any more reliable BIBREF8 . Yet, collecting the opinions of more users gives a more detailed picture of objective (or intersubjective) hatefulness. For the same reason, researchers might want to consider hate speech detection a regression problem, predicting, for example, the degree of hatefulness of a message, instead of a binary yes-or-no classification task. In the future, finding the characteristics that make users consider content hateful will be useful for building a model that automatically detects hate speech and users who spread hateful content, and for determining what makes users disseminate hateful content. ### Acknowledgments
This work was supported by the Deutsche Forschungsgemeinschaft (DFG) under grant No. GRK 2167, Research Training Group ”User-Centred Social Media”. Table 1: Summary statistics with p values and effect size estimates from WMW tests. Not all participants chose to report their age or gender. Figure 1: Reliability (Krippendorff’s a) for the different groups and questions | German |
Why does the narrator lie to his son?
A. Even though his son is a young man sooner than already, he is still too young to learn the full scope of the truth.
B. A joke stops working when someone attempts to explain it.
C. For his joke to have its desired effect, no one can know the full extent of his experiment.
D. He is an eccentric and must abide by his personal eccentricities.
| Volpla By WYMAN GUIN Illustrated by DICK FRANCIS [Transcriber's Note: This etext was produced from Galaxy Science Fiction May 1956. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed.] The only kind of gag worth pulling, I always maintained, was a cosmic one—till I learned the Cosmos has a really nasty sense of humor! There were three of them. Dozens of limp little mutants that would have sent an academic zoologist into hysterics lay there in the metabolic accelerator. But there were three of them . My heart took a great bound. I heard my daughter's running feet in the animal rooms and her rollerskates banging at her side. I closed the accelerator and walked across to the laboratory door. She twisted the knob violently, trying to hit a combination that would work. I unlocked the door, held it against her pushing and slipped out so that, for all her peering, she could see nothing. I looked down on her tolerantly. "Can't adjust your skates?" I asked again. "Daddy, I've tried and tried and I just can't turn this old key tight enough." I continued to look down on her. "Well, Dad-dee, I can't!" "Tightly enough." "What?" "You can't turn this old key tightly enough." "That's what I say -yud." "All right, wench. Sit on this chair." I got down and shoved one saddle shoe into a skate. It fitted perfectly. I strapped her ankle and pretended to use the key to tighten the clamp. Volplas at last. Three of them. Yet I had always been so sure I could create them that I had been calling them volplas for ten years. No, twelve. I glanced across the animal room to where old Nijinsky thrust his graying head from a cage. I had called them volplas since the day old Nijinsky's elongated arms and his cousin's lateral skin folds had given me the idea of a flying mutant. When Nijinsky saw me looking at him, he started a little tarantella about his cage. I smiled with nostalgia when the fifth fingers of his hands, four times as long as the others, uncurled as he spun about the cage. I turned to the fitting of my daughter's other skate. "Daddy?" "Yes?" "Mother says you are eccentric. Is that true?" "I'll speak to her about it." "Don't you know ?" "Do you understand the word?" "No." I lifted her out of the chair and stood her on her skates. "Tell your mother that I retaliate. I say she is beautiful." She skated awkwardly between the rows of cages from which mutants with brown fur and blue fur, too much and too little fur, enormously long and ridiculously short arms, stared at her with simian, canine or rodent faces. At the door to the outside, she turned perilously and waved. Again in the laboratory, I entered the metabolic accelerator and withdrew the intravenous needles from my first volplas. I carried their limp little forms out to a mattress in the lab, two girls and a boy. The accelerator had forced them almost to adulthood in less than a month. It would be several hours before they would begin to move, to learn to feed and play, perhaps to learn to fly. Meanwhile, it was clear that here was no war of dominant mutations. Modulating alleles had smoothed the freakish into a beautiful pattern. These were no monsters blasted by the dosage of radiation into crippled structures. They were lovely, perfect little creatures. My wife tried the door, too, but more subtly, as if casually touching the knob while calling. "Lunch, dear." "Be right there." She peeked too, as she had for fifteen years, but I blocked her view when I slipped out. "Come on, you old hermit. I have a buffet on the terrace." "Our daughter says I'm eccentric. Wonder how the devil she found out." "From me, of course." "But you love me just the same." "I adore you." She stretched on tiptoe and put her arms over my shoulders and kissed me. My wife did indeed have a delicious-looking buffet ready on the terrace. The maid was just setting down a warmer filled with hot hamburgers. I gave the maid a pinch and said, "Hello, baby." My wife looked at me with a puzzled smile. "What on Earth's got into you?" The maid beat it into the house. I flipped a hamburger and a slice of onion onto a plate and picked up the ketchup and said, "I've reached the dangerous age." "Oh, good heavens!" I dowsed ketchup over the hamburger, threw the onion on and closed it. I opened a bottle of beer and guzzled from it, blew out my breath and looked across the rolling hills and oak woods of our ranch to where the Pacific shimmered. I thought, "All this and three volplas, too." I wiped the back of my hand across my mouth and said aloud, "Yes, sir, the dangerous age. And, lady, I'm going to have fun." My wife sighed patiently. I walked over and put the arm that held the beer bottle around her shoulder and chucked her chin up with my other hand. The golden sun danced in her blue eyes. I watched that light in her beautiful eyes and said, "But you're the only one I'm dangerous about." I kissed her until I heard rollerskates coming across the terrace from one direction and a horse galloping toward the terrace from the other direction. "You have lovely lips," I whispered. "Thanks. Yours deserve the Good Housekeeping Seal of Approval, too." Our son reared the new palomino I had just bought him for his fourteenth birthday and yelled down, "Unhand that maiden, Burrhead, or I'll give you lead poisoning." I laughed and picked up my plate and sat down in a chair. My wife brought me a bowl of salad and I munched the hamburger and watched the boy unsaddle the horse and slap it away to the pasture. I thought, "By God, wouldn't he have a fit if he knew what I have back there in that lab! Wouldn't they all!" The boy carried the saddle up onto the terrace and dropped it. "Mom, I'd like a swim before I eat." He started undressing. "You look as though a little water might help," she agreed, sitting down next to me with her plate. The girl was yanking off her skates. "And I want one." "All right. But go in the house and put on your swim suit." "Oh, Mother . Why?" "Because, dear, I said so." The boy had already raced across the terrace and jack-knifed into the pool. The cool sound of the dive sent the girl scurrying for her suit. I looked at my wife. "What's the idea?" "She's going to be a young woman soon." "Is that any reason for wearing clothes? Look at him. He's a young man sooner than already." "Well, if you feel that way about it, they'll both have to start wearing clothes." I gulped the last of my hamburger and washed it down with the beer. "This place is going to hell," I complained. "The old man isn't allowed to pinch the maid and the kids can't go naked." I leaned toward her and smacked her cheek. "But the food and the old woman are still the best." "Say, what goes with you? You've been grinning like a happy ape ever since you came out of the lab." "I told you—" "Oh, not that again! You were dangerous at any age." I stood up and put my plate aside and bent over her. "Just the same, I'm going to have a new kind of fun." She reached up and grabbed my ear. She narrowed her eyes and put a mock grimness on her lips. "It's a joke," I assured her. "I'm going to play a tremendous joke on the whole world. I've only had the feeling once before in a small way, but I've always...." She twisted my ear and narrowed her eyes even more. "Like?" "Well, when my old man was pumping his first fortune out of some oil wells in Oklahoma, we lived down there. Outside this little town, I found a litter of flat stones that had young black-snakes under each slab. I filled a pail with them and took them into town and dumped them on the walk in front of the movie just as Theda Bara's matinee let out. The best part was that no one had seen me do it. They just couldn't understand how so many snakes got there. I learned how great it can be to stand around quietly and watch people encounter the surprise that you have prepared for them." She let go of my ear. "Is that the kind of fun you're going to have?" "Yep." She shook her head. "Did I say you are eccentric ?" I grinned. "Forgive me if I eat and run, dear. Something in the lab can't wait." The fact was that I had something more in the lab than I had bargained for. I had aimed only at a gliding mammal a little more efficient than the Dusky Glider of Australia, a marsupial. Even in the basically mutating colony, there had been a decidedly simian appearance in recent years, a long shift from the garbage-dump rats I had started with. But my first volplas were shockingly humanoid. They were also much faster than had been their predecessors in organizing their nervous activity after the slumbrous explosion of growth in the metabolic accelerator. When I returned to the lab, they were already moving about on the mattress and the male was trying to stand. He was a little the larger and stood twenty-eight inches high. Except for the face, chest and belly, they were covered with a soft, almost golden down. Where it was bare of this golden fur, the skin was pink. On their heads and across the shoulders of the male stood a shock of fur as soft as chinchilla. The faces were appealingly humanoid, except that the eyes were large and nocturnal. The cranium was in the same proportion to the body as it is in the human. When the male spread his arms, the span was forty-eight inches. I held his arms out and tried to tease the spars open. They were not new. The spars had been common to the basic colony for years and were the result of serial mutations effecting those greatly elongated fifth fingers that had first appeared in Nijinsky. No longer jointed like a finger, the spar turned backward sharply and ran alongside the wrist almost to the elbow. Powerful wrist muscles could snap it outward and forward. Suddenly, as I teased the male volpla, this happened. The spars added nine inches on each side to his span. As they swept out and forward, the lateral skin that had, till now, hung in resting folds was tightened in a golden plane that stretched from the tip of the spar to his waist and continued four inches wide down his legs to where it anchored at the little toe. This was by far the most impressive plane that had appeared till now. It was a true gliding plane, perhaps even a soaring one. I felt a thrill run along my back. By four o'clock that afternoon, I was feeding them solid food and, with the spars closed, they were holding little cups and drinking water from them in a most humanlike way. They were active, curious, playful and decidedly amorous. Their humanoid qualities were increasingly apparent. There was a lumbar curvature and buttocks. The shoulder girdle and pectoral muscles were heavy and out of proportion, of course, yet the females had only one pair of breasts. The chin and jaw were humanlike instead of simian and the dental equipment was appropriate to this structure. What this portended was brought home to me with a shock. I was kneeling on the mattress, cuffing and roughing the male as one might a puppy dog, when one of the females playfully climbed up my back. I reached around and brought her over my shoulder and sat her down. I stroked the soft fur on her head and said, "Hello, pretty one. Hello." The male watched me, grinning. He said, "'Ello, 'ello." As I walked into the kitchen, giddy with this enormous joke, my wife said, "Guy and Em are flying up for dinner. That rocket of Guy's they launched in the desert yesterday was a success. It pulled Guy up to Cloud Nine and he wants to celebrate." I danced a little jig the way old Nijinsky might do it. "Oh, great! Oh, wonderful! Good old Guy! Everybody's a success. It's great. It's wonderful. Success on success!" I danced into the kitchen table and tipped over a basket of green corn. The maid promptly left the kitchen for some other place. My wife just stared at me. "Have you been drinking the lab alcohol?" "I've been drinking the nectar of the gods. My Hera, you're properly married to Zeus. I've my own little Greeks descended from Icarus." She pretended a hopeless sag of her pretty shoulders. "Wouldn't you just settle for a worldly martini?" "I will, yes. But first a divine kiss." I sipped at my martini and lounged in a terrace chair watching the golden evening slant across the beautiful hills of our ranch. I dreamed. I would invent a euphonious set of words to match the Basic English vocabulary and teach it to them as their language. They would have their own crafts and live in small tree houses. I would teach them legends: that they had come from the stars, that they had subsequently watched the first red men and then the first white men enter these hills. When they were able to take care of themselves, I would turn them loose. There would be volpla colonies all up and down the Coast before anyone suspected. One day, somebody would see a volpla. The newspapers would laugh. Then someone authoritative would find a colony and observe them. He would conclude, "I am convinced that they have a language and speak it intelligently." The government would issue denials. Reporters would "expose the truth" and ask, "Where have these aliens come from?" The government would reluctantly admit the facts. Linguists would observe at close quarters and learn the simple volpla language. Then would come the legends. Volpla wisdom would become a cult—and of all forms of comedy, cults, I think, are the funniest. "Darling, are you listening to me?" my wife asked with impatient patience. "What? Sure. Certainly." "You didn't hear a word. You just sit there and grin into space." She got up and poured me another martini. "Here, maybe this will sober you up." I pointed. "That's probably Guy and Em." A 'copter sidled over the ridge, then came just above the oak woods toward us. Guy set it gently on the landing square and we walked down to meet them. I helped Em out and hugged her. Guy jumped out, asking, "Do you have your TV set on?" "No," I answered. "Should I?" "It's almost time for the broadcast. I was afraid we would miss it." "What broadcast?" "From the rocket." "Rocket?" "For heaven's sake, darling," my wife complained, "I told you about Guy's rocket being a success. The papers are full of it. So are the broadcasts." As we stepped up on the terrace, she turned to Guy and Em. "He's out of contact today. Thinks he's Zeus." I asked our son to wheel a TV set out onto the terrace while I made martinis for our friends. Then we sat down and drank the cocktails and the kids had fruit juice and we watched the broadcast Guy had tuned in. Some joker from Cal Tech was explaining diagrams of a multi-stage rocket. After a bit, I got up and said, "I have something out in the lab I want to check on." "Hey, wait a minute," Guy objected. "They're about to show the shots of the launching." My wife gave me a look; you know the kind. I sat down. Then I got up and poured myself another martini and freshened Em's up, too. I sat down again. The scene had changed to a desert launching site. There was old Guy himself explaining that when he pressed the button before him, the hatch on the third stage of the great rocket in the background would close and, five minutes later, the ship would fire itself. Guy, on the screen, pushed the button, and I heard Guy, beside me, give a sort of little sigh. We watched the hatch slowly close. "You look real good," I said. "A regular Space Ranger. What are you shooting at?" "Darling, will you please—be— quiet ?" "Yeah, Dad. Can it, will you? You're always gagging around." On the screen, Guy's big dead-earnest face was explaining more about the project and suddenly I realized that this was an instrument-bearing rocket they hoped to land on the Moon. It would broadcast from there. Well, now—say, that would be something! I began to feel a little ashamed of the way I had been acting and I reached out and slapped old Guy on the shoulder. For just a moment, I thought of telling him about my volplas. But only for a moment. A ball of flame appeared at the base of the rocket. Miraculously, the massive tower lifted, seemed for a moment merely to stand there on a flaming pillar, then was gone. The screen returned to a studio, where an announcer explained that the film just shown had been taken day before yesterday. Since then, the rocket's third stage was known to have landed successfully at the south shore of Mare Serenitatis. He indicated the location on a large lunar map behind him. "From this position, the telemeter known as Rocket Charlie will be broadcasting scientific data for several months. But now, ladies and gentlemen, we will clear the air for Rocket Charlie's only general broadcast. Stand by for Rocket Charlie." A chronometer appeared on the screen and, for several seconds, there was silence. I heard my boy whisper, "Uncle Guy, this is the biggest!" My wife said, "Em, I think I'll just faint." Suddenly there was a lunar landscape on the screen, looking just as it's always been pictured. A mechanical voice cut in. "This is Rocket Charlie saying, 'Hello, Earth,' from my position in Mare Serenitatis. First I will pan the Menelaus Mountains for fifteen seconds. Then I will focus my camera on Earth for five seconds." The camera began to move and the mountains marched by, stark and awesomely wild. Toward the end of the movement, the shadow of the upright third stage appeared in the foreground. Abruptly the camera made a giddy swing, focused a moment, and we were looking at Earth. At that time, there was no Moon over California. It was Africa and Europe we were looking at. "This is Rocket Charlie saying, 'Good-by, Earth.'" Well, when that screen went dead, there was pandemonium around our terrace. Big old Guy was so happy, he was wiping tears from his eyes. The women were kissing him and hugging him. Everybody was yelling at once. I used the metabolic accelerator to cut the volplas' gestation down to one week. Then I used it to bring the infants to maturity in one month. I had luck right off. Quite by accident, the majority of the early infants were females, which sped things up considerably. By the next spring, I had a colony of over a hundred volplas and I shut down the accelerator. From now on, they could have babies in their own way. I had devised the language for them, using Basic English as my model, and during the months while every female was busy in the metabolic accelerator, I taught the language to the males. They spoke it softly in high voices and the eight hundred words didn't seem to tax their little skulls a bit. My wife and the kids went down to Santa Barbara for a week and I took the opportunity to slip the oldest of the males and his two females out of the lab. I put them in the jeep beside me and drove to a secluded little valley about a mile back in the ranch. They were all three wide-eyed at the world and jabbered continuously. They kept me busy relating their words for "tree," "rock," "sky" to the objects. They had a little trouble with "sky." Until I had them out in the open country, it had been impossible to appreciate fully what lovely little creatures they were. They blended perfectly with the California landscape. Occasionally, when they raised their arms, the spars would open and spread those glorious planes. Almost two hours went by before the male made it into the air. His playful curiosity about the world had been abandoned momentarily and he was chasing one of the girls. As usual, she was anxious to be caught and stopped abruptly at the bottom of a little knoll. He probably meant to dive for her. But when he spread his arms, the spars snapped out and those golden planes sheared into the air. He sailed over her in a stunning sweep. Then he rose up and up until he hung in the breeze for a long moment, thirty feet above the ground. He turned a plaintive face back to me, dipped worriedly and skimmed straight for a thorn bush. He banked instinctively, whirled toward us in a golden flash and crashed with a bounce to the grass. The two girls reached him before I did and stroked and fussed over him so that I could not get near. Suddenly he laughed with a shrill little whoop. After that, it was a carnival. They learned quickly and brilliantly. They were not fliers; they were gliders and soarers. Before long, they took agilely to the trees and launched themselves in beautiful glides for hundreds of feet, banking, turning and spiraling to a gentle halt. I laughed out loud with anticipation. Wait till the first pair of these was brought before a sheriff! Wait till reporters from the Chronicle motored out into the hills to witness this! Of course, the volplas didn't want to return to the lab. There was a tiny stream through there and at one point it formed a sizable pool. They got into this and splashed their long arms about and they scrubbed each other. Then they got out and lay on their backs with the planes stretched to dry. I watched them affectionately and wondered about the advisability of leaving them out here. Well, it had to be done sometime. Nothing I could tell them about surviving would help them as much as a little actual surviving. I called the male over to me. He came and squatted, conference fashion, the elbows resting on the ground, the wrists crossed at his chest. He spoke first. "Before the red men came, did we live here?" "You lived in places like this all along these mountains. Now there are very few of you left. Since you have been staying at my place, you naturally have forgotten the ways of living outdoors." "We can learn again. We want to stay here." His little face was so solemn and thoughtful that I reached out and stroked the fur on his head reassuringly. We both heard the whir of wings overhead. Two mourning doves flew across the stream and landed in an oak on the opposite hillside. I pointed. "There's your food, if you can kill it." He looked at me. "How?" "I don't think you can get at them in the tree. You'll have to soar up above and catch one of them on the wing when they fly away. Think you can get up that high?" He looked around slowly at the breeze playing in the branches and dancing along the hillside grass. It was as if he had been flying a thousand years and was bringing antique wisdom to bear. "I can get up there. I can stay for a while. How long will they be in the tree?" "Chances are they won't stay long. Keep your eye on the tree in case they leave while you are climbing." He ran to a nearby oak and clambered aloft. Presently he launched himself, streaked down-valley a way and caught a warm updraft on a hillside. In no time, he was up about two hundred feet. He began criss-crossing the ridge, working his way back to us. The two girls were watching him intently. They came over to me wonderingly, stopping now and then to watch him. When they were standing beside me, they said nothing. They shaded their eyes with tiny hands and watched him as he passed directly above us at about two hundred and fifty feet. One of the girls, with her eyes fast on his soaring planes, reached out and grasped my sleeve tightly. He flashed high above the stream and hung behind the crest of the hill where the doves rested. I heard their mourning from the oak tree. It occurred to me they would not leave that safety while the hawklike silhouette of the volpla marred the sky so near. I took the girl's hand from my sleeve and spoke to her, pointing as I did so. "He is going to catch a bird. The bird is in that tree. You can make the bird fly so that he can catch it. Look here." I got up and found a stick. "Can you do this?" I threw the stick up into a tree near us. Then I found her a stick. She threw it better than I had expected. "Good, pretty one. Now run across the stream and up to that tree and throw a stick into it." She climbed skillfully into the tree beside us and launched herself across the stream. She swooped up the opposite hillside and landed neatly in the tree where the doves rested. The birds came out of the tree, climbing hard with their graceful strokes. I looked back, as did the girl remaining beside me. The soaring volpla half closed his planes and started dropping. He became a golden flash across the sky. The doves abruptly gave up their hard climbing and fell away with swiftly beating wings. I saw one of the male volpla's planes open a little. He veered giddily in the new direction and again dropped like a molten arrow. The doves separated and began to zigzag down the valley. The volpla did something I would not have anticipated—he opened his planes and shot lower than the bird he was after, then swept up and intercepted the bird's crossward flight. I saw the planes close momentarily. Then they opened again and the bird plummeted to a hillside. The volpla landed gently atop the hill and stood looking back at us. The volpla beside me danced up and down shrieking in a language all her own. The girl who had raised the birds from the tree volplaned back to us, yammering like a bluejay. It was a hero's welcome. He had to walk back, of course—he had no way to carry such a load in flight. The girls glided out to meet him. Their lavish affection held him up for a time, but eventually he strutted in like every human hunter. They were raptly curious about the bird. They poked at it, marveled at its feathers and danced about it in an embryonic rite of the hunt. But presently the male turned to me. "We eat this?" I laughed and took his tiny, four-fingered hand. In a sandy spot beneath a great tree that overhung the creek, I built a small fire for them. This was another marvel, but first I wanted to teach them how to clean the bird. I showed them how to spit it and turn it over their fire. Later, I shared a small piece of the meat in their feast. They were gleeful and greasily amorous during the meal. When I had to leave, it was dark. I warned them to stand watches, keep the fire burning low and take to the tree above if anything approached. The male walked a little away with me when I left the fire. I said again, "Promise me you won't leave here until we've made you ready for it." "We like it here. We will stay. Tomorrow you bring more of us?" "Yes. I will bring many more of you, if you promise to keep them all here in this woods until they're ready to leave." "I promise." He looked up at the night sky and, in the firelight, I saw his wonder. "You say we came from there?" "The old ones of your kind told me so. Didn't they tell you?" "I can't remember any old ones. You tell me." "The old ones told me you came long before the red men in a ship from the stars." Standing there in the dark, I had to grin, visioning the Sunday supplements that would be written in about a year, maybe even less. He looked into the sky for a long time. "Those little lights are the stars?" "That's right." "Which star?" I glanced about and presently pointed over a tree. "From Venus." Then I realized I had blundered by passing him an English name. "In your language, Pohtah." He looked at the planet a long time and murmured, "Venus. Pohtah." That next week, I transported all of the volplas out to the oak woods. There were a hundred and seven men, women and children. With no design on my part, they tended to segregate into groups consisting of four to eight couples together with the current children of the women. Within these groups, the adults were promiscuous, but apparently not outside the group. The group thus had the appearance of a super-family and the males indulged and cared for all the children without reference to actual parenthood. By the end of the week, these super-families were scattered over about four square miles of the ranch. They had found a new delicacy, sparrows, and hunted them easily as they roosted at night. I had taught the volplas to use the fire drill and they were already utilizing the local grasses, vines and brush to build marvelously contrived tree houses in which the young, and sometimes the adults, slept through midday and midnight. The afternoon my family returned home, I had a crew of workmen out tearing down the animal rooms and lab building. The caretakers had anesthetized all the experimental mutants, and the metabolic accelerator and other lab equipment was being dismantled. I wanted nothing around that might connect the sudden appearance of the volplas with my property. It was already apparent that it would take the volplas only a few more weeks to learn their means of survival and develop an embryonic culture of their own. Then they could leave my ranch and the fun would be on. My wife got out of the car and looked around at the workmen hurrying about the disemboweled buildings and she said, "What on Earth is going on here?" "I've finished my work and we no longer need the buildings. I'm going to write a paper about my results." My wife looked at me appraisingly and shook her head. "I thought you meant it. But you really ought to. It would be your first." My son asked, "What happened to the animals?" "Turned them over to the university for further study," I lied. "Well," he said to her, "you can't say our pop isn't a man of decision." Twenty-four hours later, there wasn't a sign of animal experimentation on the ranch. Except, of course, that the woods were full of volplas. At night, I could hear them faintly when I sat out on the terrace. As they passed through the dark overhead, they chattered and laughed and sometimes moaned in winged love. One night a flight of them soared slowly across the face of the full Moon, but I was the only one who noticed. | C. For his joke to have its desired effect, no one can know the full extent of his experiment. |
Which region had the worst topline performance for MGM during FY2022? | Evidence 0:
Las Vegas Strip Resorts
Net revenues of $8.4 billion in the current year compared to $4.7 billion in the prior year, an
increase of 77%;
Evidence 1:
Regional Operations
Net revenues of $3.8 billion in the current year compared to $3.4 billion in the prior year, an
increase of 12%;
Evidence 2:
MGM China
Net revenues of $674 million in the current year compared to $1.2 billion in the prior year, a
decrease of 44%; | MGM China experienced the worst topline performance amongst the other regions presented. Its revenue declined 44% in FY2022 whereas the other regions presented increased their revenues. |
What was Captain Walsh's main motive behind putting the narrator on the mission?
A. Walsh sought revenge against the narrator.
B. Walsh wanted to test the narrator's intelligence.
C. Walsh wanted the narrator fired from his position.
D. Walsh wanted to test the narrator's competency.
| A PLANET NAMED JOE By S. A. LOMBINO There were more Joes on Venus than you could shake a ray-gun at. Perhaps there was method in Colonel Walsh's madness—murder-madness—when he ordered Major Polk to scan the planet for a guy named Joe. [Transcriber's Note: This etext was produced from Planet Stories November 1952. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed.] Colonel Walsh had a great sense of humor. I hated his guts ever since we went through the Academy together, but he had a great sense of humor. For example, he could have chosen a Second Looie for the job on Venus. He might even have picked a Captain. But he liked me about as much as I liked him, and so he decided the job was just right for a Major. At least, that's what he told me. I stood at attention before his desk in the Patrol Station. We were somewhere in Area Two on Earth, takeoff point for any operations in Space II. The duty was fine, and I liked it a lot. Come to think of it, the most I ever did was inspect a few defective tubes every now and then. The rest was gravy, and Colonel Walsh wasn't going to let me get by with gravy. "It will be a simple assignment, Major," he said to me, peering over his fingers. He held them up in front of him like a cathedral. "Yes, sir," I said. "It will involve finding one man, a Venusian native." I wanted to say, "Then why the hell don't you send a green kid on the job? Why me?" Instead, I nodded and watched him playing with his fingers. "The man is a trader of sorts. Rather intelligent." He paused, then added, "For a native, that is." I had never liked Walsh's attitude toward natives. I hadn't liked the way he'd treated the natives on Mars ever since he'd taken over there. Which brought to mind an important point. "I always figured Venus was under the jurisdiction of Space III, sir. I thought our activities were confined to Mars." He folded his fingers like a deck of cards and dropped them on his desk as if he were waiting for me to cut. "Mmmm," he said, "yes, that's true. But this is a special job. It so happens this Venusian is the one man who can help us understand just what's happening on Mars." I tried to picture a Venusian understanding Mars and I didn't get very far. "He's had many dealings with the natives there," Walsh explained. "If anyone can tell us the reasons for the revolt, he can." If Walsh really wanted to know the reasons for the revolt, I could give them to him in one word: Walsh. I had to laugh at the way he called it "revolt." It had been going on for six months now and we'd lost at least a thousand men from Space II. Revolt. "And this man is on Venus now?" I asked for confirmation. I'd never been to Venus, being in Space II ever since I'd left the Moon run. It was just like Walsh to ship me off to a strange place. "Yes, Major," he said. "This man is on Venus." At the Academy he had called me Fred. That was before I'd reported him for sleeping on Boiler Watch. He'd goofed off on a pile of uranium that could've, and almost did, blow the barracks sky-high that night. He still thought it was my fault, as if I'd done the wrong thing by reporting him. And now, through the fouled-up machinery that exists in any military organization, he outranked me. "And the man's name, sir?" "Joe." A tight smile played on his face. "Joe what?" I asked. "Just Joe." "Just Joe?" "Yes," Walsh said. "A native, you know. They rarely go in for more than first names. But then, it should be simple to find a man with a name like Joe. Among the natives, I mean." "I don't know, sir." "A relatively simple assignment," Walsh said. "Can you tell me anything else about this man? Physical appearance? Personal habits? Anything?" Walsh seemed to consider this for a moment. "Well, physically he's like any of the other Venusians, so I can't give you much help there. He does have a peculiar habit, though." "What's that?" "He has an affinity for Terran cigarettes." I sighed. "Well, it's not very much to go on." "You'll find him," Walsh said, grinning. "I'm sure of it." The trip to Venus came off without a hitch. I did a lot of thinking on that trip. I thought about Mars and the revolt there. And I thought about Colonel Leonard Walsh and how he was supposed to be quelling that revolt. Ever since Walsh had taken command, ever since he'd started pushing the natives around, there'd been trouble. It was almost as if the whole damned planet had blown up in our faces the moment he took over. Swell guy, Walsh. Venus was hotter than I'd expected it to be. Much too hot for the tunic I was wearing. It smelled, too. A funny smell I couldn't place. Like a mixture of old shoe and after-shave. There were plants everywhere I looked. Big plants and small ones, some blooming with flowers I'd never seen before, and some as bare as cactus. I recognized a blue figure as one of the natives the pilot had told me about. He was tall, looking almost human except that everything about him was elongated. His features, his muscles, everything seemed to have been stretched like a rubber band. I kept expecting him to pop back to normal. Instead, he flashed a double row of brilliant teeth at me. I wondered if he spoke English. "Hey, boy," I called. He ambled over with long-legged strides that closed the distance between us in seconds. "Call me Joe," he said. I dropped my bags and stared at him. Maybe this was going to be a simple assignment after all. "I sure am glad to see you, Joe," I said. "Same here, Toots," he answered. "The guys back in Space II are searching high and low for you," I told him. "You've got the wrong number," he said, and I was a little surprised at his use of Terran idiom. "You are Joe, aren't you? Joe the trader?" "I'm Joe, all right," he said. "Only thing I ever traded, though, was a pocketknife. Got a set of keys for it." "Oh," I said, my voice conveying my disappointment. I sighed and began wondering just how I should go about contacting the Joe I was looking for. My orders said I was to report to Captain Bransten immediately upon arrival. I figured the hell with Captain Bransten. I outranked him anyway, and there wasn't much he could do if I decided to stop for a drink first. "Where's the Officer's Club?" I asked the Venusian. "Are you buying information or are you just curious?" "Can you take me there?" I asked. "Sure thing, Toots." He picked up my bags and started walking up a heavily overgrown path. We'd probably walked for about ten minutes when he dropped my bags and said, "There it is." The Officer's Club was a plasteel hut with window shields that protected it from the heat of the sun. It didn't look too comfortable but I really wanted that drink. I reached into my tunic and slipped the native thirty solars. He stared at the credits curiously and then shrugged his shoulders. "Oh well, you're new here. We'll let it go." He took off then, while I stared after him, wondering just what he'd meant. Had I tipped him too little? I shrugged and looked over at the Officer's Club. From the outside it looked as hot as hell. On the inside it was about two degrees short of that mark. I began to curse Walsh for taking me away from my nice soft job in Space II. There wasn't much inside the club. A few tables and chairs, a dart game and a bar. Behind the bar a tall Venusian lounged. I walked over and asked, "What are you serving, pal?" "Call me Joe," he answered. He caught me off balance. "What?" "Joe," he said again. A faint glimmer of understanding began to penetrate my thick skull. "You wouldn't happen to be Joe the trader? The guy who knows all about Mars, would you?" "I never left home," he said simply. "What are you drinking?" That rat! That dirty, filthy, stinking, unprincipled.... But then, it should be simple to find a man with a name like Joe. Among the natives, I mean. Sure. Oh sure. Real simple. Walsh was about the lowest, most contemptible.... "What are you drinking, pal?" the Venusian asked again. "Skip it," I said. "How do I get to the captain's shack?" "Follow your nose, pal. Can't miss it." I started to pick up my bag as another Venusian entered. He waved at the bartender. "Hello, Joe," he said. "How's it going?" "Not so hot, Joe," the bartender replied. I listened in fascination. Joe, Joe, Joe. So this was Walsh's idea of a great gag. Very funny. Very.... "You Major Polk, sweetheart?" the Venusian who'd just come in asked. "Yes," I said, still thinking of Colonel Walsh. "You better get your butt over to the captain's shack," he said. "He's about ready to post you as overdue." "Sure," I said wearily. "Will you take my bags, please?" "Roger," he answered. He picked up the bags and nodded at the bar. "So long, Joe," he said to the bartender. "See you, Joe," the bartender called back. Captain Bransten was a mousey, unimpressive sort of man. He was wearing a tropical tunic, but he still resembled a wilted lily more than he did an officer. "Have a seat, Major," he offered. He reached for a cigarette box on the desk and extended it to me. He coughed in embarrassment when he saw it was empty. Quickly, he pressed a button on his desk and the door popped open. A tall, blue Venusian stepped lithely into the room. "Sir?" the Venusian asked. "We're out of cigarettes, Joe," the Captain said. "Will you get us some, please?" "Sure thing," the Venusian answered. He smiled broadly and closed the door behind him. Another Joe , I thought. Another damned Joe. "They steal them," Captain Bransten said abruptly. "Steal what?" I asked. "Cigarettes. I sometimes think the cigarette is one of the few things they like about Terran culture." So Walsh had taken care of that angle too. He does have a peculiar habit, though. He has an affinity for Terran cigarettes. Cigarettes was the tip I should have given; not solars. "All right," I said, "suppose we start at the beginning." Captain Bransten opened his eyes wide. "Sir?" he asked. "What's with all this Joe business? It may be a very original name but I think its popularity here is a little outstanding." Captain Bransten began to chuckle softly. I personally didn't think it was so funny. I tossed him my withering Superior Officer's gaze and waited for his explanation. "I hadn't realized this was your first time on Venus," he said. "Is there a local hero named Joe?" I asked. "No, no, nothing like that," he assured me. "It's a simple culture, you know. Not nearly as developed as Mars." "I can see that," I said bitingly. "And the natives are only now becoming acquainted with Terran culture. Lots of enlisted men, you know." I began to get the idea. And I began to appreciate Walsh's doubtful ancestry more keenly. "It's impossible to tell exactly where it all started, of course," Bransten was saying. I was beginning to get angry. Very angry. I was thinking of Walsh sitting back in a nice cozy foam chair back on Earth. "Get to the point, Captain!" I barked. "Easy, sir," Bransten said, turning pale. I could see that the Captain wasn't used to entertaining Majors. "The enlisted men. You know how they are. They'll ask a native to do something and they'll call him Joe. 'Hey, Joe, give me a hand with this.' Or 'Listen, Joe, how'd you like to earn some cigarettes?' Do you follow?" "I follow, all right," I said bitterly. "Well," Bransten went on, "that sort of thing mushrooms. The natives are a simple, almost childish people. It appealed to them—the Joe business, I mean. Now they're all Joe. They like it. That and the cigarettes." He cleared his throat and looked at me apologetically as if he were personally responsible for Venusian culture. In fact, he looked as if he were responsible for having put Venus in the heavens in the first place. "Do you understand, Major? Just a case of extended idiom, that's all." Just a case of extended idiot , I thought. An idiot on a wild goose chase a hell of a long way from home. "I understand perfectly," I snapped. "Where are my quarters?" Bransten asked a Venusian named Joe to show me my quarters, reminding me that chow was at thirteen hundred. As I was leaving, the first Venusian came back with the cigarettes Bransten had ordered. I could tell by the look on his face that he probably had half a carton stuffed into his pockets. I shrugged and went to change into a tropical tunic. I called Earth right after chow. The Captain assured me that this sort of thing was definitely against regulations, but he submitted when I twinkled my little gold leaf under his nose. Walsh's face appeared on the screen. He was smiling, looking like a fat pussy cat. "What is it, Major?" he asked. "This man Joe," I said. "Can you give me any more on him?" Walsh's grin grew wider. "Why, Major," he said, "you're not having any difficulties, are you?" "None at all," I snapped back. "I just thought I'd be able to find him a lot sooner if...." "Take your time, Major," Walsh beamed. "There's no rush at all." "I thought...." "I'm sure you can do the job," Walsh cut in. "I wouldn't have sent you otherwise." Hell, I was through kidding around. "Look...." "He's somewhere in the jungle, you know," Walsh said. I wanted to ram my fist into the screen, right smack up against those big white teeth. Instead, I cut off the transmission and watched the surprised look on his face as his screen went blank millions of miles away. He blinked at the screen, trying to realize I'd deliberately hung up on him. "Polk!" he shouted, "can you hear me?" I smiled, saw the twisted hatred on his features, and then the screen on my end went blank, too. He's somewhere in the jungle, you know. I thanked Captain Bransten for his hospitality and went back to my quarters. As I saw it, there were two courses for me to follow. One: I could say the hell with Walsh and Venus. That would mean hopping the next ship back to Earth. It would also mean disobeying the direct order of a superior officer. It might mean demotion, and it might mean getting bounced out of the Service altogether. Two: I could assume there really was a guy name Joe somewhere in that jungle, a Joe separate and apart from the other Joes on this planet, a trader Joe who knew the Martians well. I could always admit failure, of course, and return empty handed. Mission not accomplished. Or, I might really find a guy who was trader Joe. I made my decision quickly. I wanted to stay in the Service, and besides Walsh may have been on the level for the first time in his life. Maybe there was a Joe here who could help us on Mars. If there was I'd try to find him. It was still a hell of a trick though. I cursed Walsh again and pushed the buzzer near my bed. A tall Venusian stepped into the room. "Joe?" I asked, just to be sure. "Who else, boss?" he answered. "I'm trying to locate someone," I said. "I'll need a guide to take me into the jungle. Can you get me one?" "It'll cost you, boss," the Venusian said. "How much?" "Two cartons of cigarettes at least." "Who's the guide?" I asked. "How's the price sound?" "Fine, fine," I said impatiently. And the Captain had said they were almost a childish people! "His name is Joe," the Venusian told me. "Best damn guide on the planet. Take you anywhere you want to go, do anything you want to do. Courageous. Doesn't know the meaning of fear. I've known him to...." "Skip it," I said, cutting the promotion short. "Tell him to show up around fifteen hundred with a complete list of what we'll need." The Venusian started to leave. "And Joe," I said, stopping him at the door, "I hope you're not overlooking your commission on the deal." His face broke into a wide grin. "No danger of that, boss," he said. When he was gone I began figuring out a plan of action. Obviously, I'd just have to traipse through the jungle looking for a guy named Joe on a planet where everyone was named Joe. Everybody, at least, but the Captain, the small garrison attached to the Station, and me. I began wondering why Walsh had gone to so much trouble to get rid of me. The job, as I saw it, would take a hell of a long time. It seemed like a silly thing to do, just to get even with a guy for something that had happened years ago. He surely must have realized that I'd be back again, sooner or later. Maybe he had another little junket all set for me. Or maybe he didn't expect me to come back. The thought hadn't occurred to me before this, and I began to consider it seriously. Walsh was no good, rotten clear through. He was failing at the job of keeping Mars in hand, and he probably realized that a few more mistakes on his part would mean the end of his career with Space II. I chuckled as I thought of him isolated in some God-forsaken place like Space V or Space VII. This probably bothered him a lot, too. But what probably bothered him more was the fact that I was next in command. If he were transferred, I'd be in charge of Space II, and I could understand how much that would appeal to Walsh. I tried to figure the thing out sensibly, tried to weigh his good points against his bad. But it all came back to the same thing. A guy who would deliberately go to sleep on Boiler Watch with a ton of uranium ready to blast a barracks to smithereens if it wasn't watched, would deliberately do just about anything. Sending me off on a wild goose chase after a character named Joe may have been a gag. But it may have been something a little grimmer than a gag, and I made up my mind to be extremely careful from here on in. The guide arrived at fifteen hundred on the dot. He was tall, elongated, looked almost like all the other Venusians I'd seen so far. "I understand you need a Grade A guide, sir," he said. "Are you familiar with the jungle?" I asked him. "Born and raised there, sir. Know it like the back of my hand." "Has Joe told you what the payment will be?" "Yes, sir. A carton and a half of cigarettes." I thought about Joe deducting his commission and smiled. "When can we leave?" "Right away, sir. We won't need much really. I've made a list of supplies and I can get them in less than an hour. I suggest you wear light clothing, boots, and a hat." "Will I need a weapon?" He looked at me, his eyes faintly amused. "Why, what for, sir?" "Never mind," I said. "What's your name, by the way?" He lifted his eyebrows, and his eyes widened in his narrow face. He was definitely surprised. "Joe," he said. "Didn't you know?" When we'd been out for a while I discovered why Joe had suggested the boots and the hat. The undergrowth was often sharp and jagged and it would have sliced my legs to ribbons were they not protected by the high boots. The hat kept the strong sun off my head. Joe was an excellent guide and a pleasant companion. He seemed to be enjoying a great romp, seemed to love the jungle and take a secret pleasure in the work he was doing. There were times when I couldn't see three feet ahead of me. He'd stand stock still for a few minutes, his head barely moving, his eyes darting from one plant to another. Then he'd say, "This way," and take off into what looked like more impenetrable jungle invariably to find a little path leading directly to another village. Each village was the same. The natives would come running out of their huts, tall and blue, shouting, "Cigarettes, Joe? Cigarettes?" It took me a while to realize they were addressing me and not my guide. Everybody was Joe. It was one beautiful, happy, joyous round of stinking, hot jungle. And I wasn't getting any nearer my man. Nor had I any idea how I was supposed to find him. I began to feel pretty low about the whole affair. Joe, on the other hand, enjoyed every moment of the trip. In each village he greeted the natives cheerfully, told them stories, swapped gossip and jokes. And when it was time to leave, he would say goodbye to all his friends and we would plunge into the twisted foliage again. His spirits were always high and he never failed to say the right thing that would give a momentary lift to my own depressed state of mind. He would talk for hours on end as we hacked our way through the jungle. "I like Venus," he said once. "I would never leave it." "Have you ever been to Earth?" I asked. "No," Joe replied. "I like Terrans too, you understand. They are good for Venus. And they are fun." "Fun?" I asked, thinking of a particular species of Terran: species Leonard Walsh. "Yes, yes," he said wholeheartedly. "They joke and they laugh and ... well, you know." "I suppose so," I admitted. Joe smiled secretly, and we pushed on. I began to find, more and more, that I had started to talk freely to Joe. In the beginning he had been just my guide. There had been the strained relationship of employer and employee. But as the days lengthened into weeks, the formal atmosphere began to crumble. I found myself telling him all about Earth, about the people there, about my decision to attend the Academy, the rigid tests, the grind, even the Moon run. Joe was a good listener, nodding sympathetically, finding experiences in his own life to parallel my own. And as our relationship progressed from a casual one to a definitely friendly one, Joe seemed more enthusiastic than ever to keep up our grinding pace to find what we were looking for. Once we stopped in a clearing to rest. Joe lounged on the matted greenery, his long body stretched out in front of him, the knife gleaming in his belt. I'd seen him slash his way through thick, tangled vines with that knife, his long, muscular arms powerfully slicing through them like strips of silk. "How far are we from the Station?" I asked. "Three or four Earth weeks," he replied. I sighed wearily. "Where do we go from here?" "There are more villages," he said. "We'll never find him." "Possibly," Joe mused, the smile creeping over his face again. "A wild goose chase. A fool's errand." "We'd better get started," Joe said simply. I got to my feet and we started the march again. Joe was still fresh, a brilliant contrast to me, weary and dejected. Somehow, I had the same feeling I'd had a long time ago on my sixteenth birthday. One of my friends had taken me all over the city, finally dropping me off at my own house where the whole gang was gathered for a surprise party. Joe reminded me of that friend. "There's a village ahead," he said, and the grin on his face was large now, his eyes shining. Something was missing here. Natives. There were no natives rushing out to greet us. No cries of "Cigarettes? Cigarettes?" I caught up with Joe. "What's the story?" I whispered. He shrugged knowingly and continued walking. And then I saw the ship, nose pointing into space, catching the rays of the sun like a great silver bullet. "What...?" I started. "It's all right," Joe said, smiling. The ship looked vaguely familiar. I noticed the crest of Space II near the nose, and a lot of things became clear then. I also saw Walsh standing near one of the huts, a stun gun in his hand. "Hello, Major," he called, almost cheerfully. The gun didn't look cheerful, though. It was pointed at my head. "Fancy meeting you here, Colonel," I said, trying to match his joviality. Somehow it didn't quite come off. Joe was walking beside me, waving at the colonel, beaming all over with happiness. "I see you found your man," Walsh said. I turned rapidly. Joe nodded and kept grinning, a grin that told me he was getting a big kick out of all this. Like a kid playing a game. I faced Walsh again. "Okay, what's it all about, pal?" "Colonel," Walsh corrected me. "You mustn't forget to say Colonel, Major ." He emphasized my rank, and he said it with a sort of ruthless finality. I waited. I could see he was just busting to tell me how clever he'd been. Besides, there wasn't much I could do but wait. Not with Walsh pointing the stun gun at my middle. "We've come a long way since the Academy, haven't we, Major?" "If you mean in miles," I said, looking around at the plants, "we sure have." Walsh grinned a little. "Always the wit," he said drily. And then the smile faded from his lips and his eyes took on a hard lustre. "I'm going to kill you, you know." He said it as if he were saying, "I think it'll rain tomorrow." Joe almost clapped his hands together with glee. He was really enjoying this. Another of those funny Terran games. "You gave me a powerful handicap to overcome," Walsh said. "I suppose I should thank you, really." "You're welcome," I said. "It wasn't easy living down the disgrace you caused me." "It was your own damn fault," I said. "You knew what you were doing when you decided to cork off." Beside me, Joe chuckled a little, enjoying the game immensely. "You didn't have to report me," Walsh said. "No? Maybe I should have forgotten all about it? Maybe I should have nudged you and served you orange juice? So you could do it again sometime and maybe blow up the whole damn Academy!" Walsh was silent for a long time. When he spoke his voice was barely audible. The heat was oppressive, as if it were concentrated on this little spot in the jungle, focusing all its penetration on a small, unimportant drama. I could hear Joe breathing beside me. "I'm on my way out," Walsh rasped. "Finished, do you understand?" "Good," I said. And I meant it. "This Mars thing. A terrible fix. Terrible." Beside me, a slight frown crossed Joe's face. Apparently he couldn't understand the seriousness of our voices. What had happened to the game, the fun? "You brought the Mars business on yourself," I told Walsh. "There was never any trouble before you took command." "The natives," he practically shouted. "They ... they...." Joe caught his breath sharply, and I wondered what Walsh was going to say about the natives. Apparently he'd realized that Joe was a native. Or maybe Joe's knife had something to do with it. "What about the natives?" I asked. "Nothing," Walsh said. "Nothing." He was silent for a while. "A man of my calibre," he said then, his face grim. "Dealing with savages." He caught himself again and threw a hasty glance at Joe. The perplexed frown had grown heavier on Joe's face. He looked at the colonel in puzzlement. | A. Walsh sought revenge against the narrator. |
How does their perturbation algorihm work? | ### Introduction
At the core of Natural Language Processing (NLP) neural models are pre-trained word embeddings like Word2Vec BIBREF0, GloVe BIBREF1 and ELMo BIBREF2. They help initialize the neural models, lead to faster convergence and have improved performance for numerous application such as Question Answering BIBREF3, Summarization BIBREF4, Sentiment Analysis BIBREF5. While word embeddings are powerful in unlimited constraints such as computation power and compute resources, it becomes challenging to deploy them to on-device due to their huge size. This led to interesting research by BIBREF6, BIBREF7, BIBREF8, who showed that actually word embedding can be replaced with lightweight binary LSH projections learned on-the-fly. The projection approach BIBREF9, BIBREF10 surmounts the need to store any embedding matrices, since the projections are dynamically computed. This further enables user privacy by performing inference directly on device without sending user data (e.g., personal information) to the server. The computation of the representation is linear in the number of inputs in the sentence surmounting the need to maintain and lookup global vocabulary, and reducing the memory size to $O(|T \cdot d|)$. The projection representations can operate on word and character level, and can be used to represent a sentence or a word depending on the NLP application. BIBREF6 have shown that on-device LSH projections lead to state-of-the-art results in dialog act classification and reach significant improvement upon prior LSTM and CNN neural models. Despite being so successful, yet there are no studies showing the properties and power of LSH projections. In this paper, we address that by studying What makes projection models effective? and Are these projection models resistant to perturbations and misspellings in input text? To answer these questions, we conduct a series of experimental studies and analysis. For instance, by studying the collision of the learned projection representations, we verify the effectiveness of the produced representations. Our study showed that LSH projections have low collision, meaning that the representations are good allowing the model to capture the meaning of words, instead of colliding everything into one meaning. Next, by analyzing the different character perturbations, we show the robustness of LSH projections when modeling word or sentence level representations. The intuition is that the projection should be able to capture word misspellings as similar, and yet it should be robust to semantically dissimilar terms. We show that Self-Governing Neural Networks (SGNN) models BIBREF6 evaluated with perturbed LSH projections are resistant to misspellings and transformation attacks, while LSTMs with increased perturbations dropped in performance. Overall, the studies are very interesting showcasing the robustness of LSH projection representations, their resistance to misspellings and transformations, and also explains why they lead to better performance. ### Background: LSH projections for text representations
The Projection function, $\mathbb {P}$ (Figure FIGREF1), BIBREF9 used in SGNN models BIBREF6 extracts token (or character) n-gram & skip-gram features from a raw input text, $\textbf {x}$ and dynamically generates a binary projection representation, $\mathbb {P}(\mathbf {x}) \in [0,1]^{T.d}$ after a Locality-Sensitive Hashing (LSH) based transformation, $\mathbb {L}$ as in where $\mathbb {F}$ extracts n-grams(or skip-grams), $[f_1, \cdots , f_n]$ from the input text. Here, $[f_1, \cdots , f_n]$ could refer to either character level or token level n-grams(or skip-grams) features. ### Collision Study
Before diving into the actual collision studies, it is important to understand what the properties of good projections are. For instance, good projections should be as separate as possible, while still capturing the inherent n-gram features. Words with similar character n-gram feature vectors should be closer to each other i.e. cat and cats, but yet separate from each other so that the network can learn that cat and cats are related, but yet different. Such observations are not evident from the projections. One way to understand them is by looking at the collision rates. For instance, if there are too many projection collisions, this means that the network is fundamentally incapable of learning and it will not be able to generalize. For the purpose, we test how spread out the projections are for word and sentence representations. We take a large corpus enwik9 and analyze the average hamming distance of the words and sentences in the corpus. Intuitively, good projections should have less collisions. Our study shows that there is almost no collision. On an average the Hamming distances between words are 557 bits, which is around 50% of the projection dimension. Standard deviations are one order of magnitude lower compared to the average Hamming distances between words which means that on average projections are more or less spread out. For high deviation, it means too many words are either too close to each other or too far away from other other. To understand the properties of word and sentence projections, we conduct two experiments, one in which we compute the word projections and another one in which we compute the sentence projections. For our experiments, we fix the projection dimension, $dim(\mathbb {P}(w)) = 1120$ ($T=80, \, d=14$) following BIBREF6. Results are shown in Table TABREF3 and Table TABREF4 respectively. Table TABREF3 shows the collision results of the word level projections. On the left we list different projection configurations by varying the number of projection functions $T$, the dimensionality $d$, turning on or off character level projections, including varying size of n-gram and skip-gram features. For each projection configuration, we show the average Hamming distance and the standard deviation. As it can be seen, by increasing the number of n-gram and skip-gram features, the words become more spread out with lesser standard deviation. We recommend using higher number of n-gram and skip-gram features for better model performance. Table TABREF4 shows the collision results of the sentence level projections. Similarly to Table TABREF3 the left side shows the different projection configurations. For each configuration, we show the average Hamming distance and standard deviation. In the sentence level projection study, we observe that when we consider only word level features, the projections are insensitive to sentence length. But with the character projections on, they are sensitive to the sentence length. This happens because the character projection space is smaller than the words space, as we see only fewer variations for the sentence projections with n-gram and skip-gram compared to word level. In sentence level projection with word level features, the dimensionality of the spacer vector is high, hence applying projections on this leads to discriminative representations. More concretely, this means that projections with large feature spaces are able to capture the distinctions between any two observed pairs and adding more words to the sentence is not going to change that. On the other hand for short sentences with character level features, the number of possible observed unique char ngrams vs those observed in longer sentences can differ. ### Perturbation Study
To further test the robustness of the projections, we conduct perturbation study. A good projection should separate out perturbed word like baank from cats. Meaning that the average Hamming distance from the collision study should be greater than the Hamming distance with and without perturbations. ### Perturbation Study ::: Character & Word Perturbations
In this section, we analyze the Hamming distance between the projections of the sentences from the enwik9 dataset and the corresponding projections of the same sentences after applying character level perturbations. We experiment with three types of character level perturbation BIBREF11 and two types of word level perturbation operations. ### Perturbation Study ::: Character Level Perturbation Operations
insert(word, n) : We randomly choose n characters from the character vocabulary and insert them at random locations into the input word. We however retain the first and last characters of the word as is. Ex. transformation: $sample \rightarrow samnple$. swap(word, n): We randomly swap the location of two characters in the word n times. As with the insert operation, we retain the first and last characters of the word as is and only apply the swap operation to the remaining characters. Ex. transformation: $sample \rightarrow sapmle$. duplicate(word, n): We randomly duplicate a character in the word by n times. Ex. transformation: $sample \rightarrow saample$. ### Perturbation Study ::: Character Level Perturbation Operations ::: Word Level Perturbation Operations
drop(sentence, n): We randomly drop n words from the sentence. Ex. transformation: This is a big cat. $\rightarrow $ This is a cat. duplicate(sentence, n): Similar to duplicate(word, n) above, we randomly duplicate a word in the sentence n times. Ex. transformation: This is a big cat. $\rightarrow $ This is a big big cat. swap(sentence, n): Similar to swap(word, n), we randomly swap the location of two words in the sentence n times. Ex. transformation: This is a big cat. $\rightarrow $ This cat is big. For both character and word level perturbations, we decide whether or not to perturb each word in a sentence with a fixed probability. For the character level perturbations, once a word is chosen for perturbation, we randomly pick one of the perturbation operations from {insert, swap, duplicate} and randomly pick the number of characters to transform $n \in \lbrace 1,\;3\rbrace $. For the word level perturbations, we randomly apply one of the operations from {drop, duplicate, swap}. We consider perturbation probabilities of $0.05$ and $0.1$ for our experiments. ### Perturbation Study ::: Discussion
We show results on multiple perturbation studies. For instance, sentence has word and character level perturbations, while word has character only perturbation. We evaluate the impact of the word and character projections for sentence and word level projections on the enwik9 dataset. Table TABREF13 shows the character and word perturbation with sentence level projections. Table TABREF14 shows the character perturbation for word level projections. We observe that the hamming distances between the projections of the perturbed versions of the same words are significantly smaller than the average distance of the word projections measured in the collision study in Section SECREF3. This shows that the words are well separated in the projection space and could potentially be less susceptible to misspellings and omissions. Based on the results in all Tables 1 to 4, we found a nice linear relationship between the hamming distance, the projection dimension and the amount of perturbation. As it can be seen in the results, the hamming distance between the projections before and after perturbation is directly proportional to the product of the projection dimension and percentage of perturbation as follows: $ \Delta _{\mathbb {P}_{m}} = K_{m}\, \cdot T \, \cdot \, d \cdot P_{perturb} \; , m \in \lbrace word, \,character\rbrace , \; K_{m} > 0$ where $\Delta _{\mathbb {P}_{m}}$ refers to the hamming distance between the projections before and after perturbations and $m$ refers to the mode of projection - {word, character}. $T \cdot d$ refers to the projection space dimension and $P_{perturb}$ refers to the probability of perturbation. $K_{m} > 0$ is a proportionality constant which depends on the projection mode. We observe that $K_{word} > K_{char}$ from our experiments. Character mode projections are relatively more robust to perturbations, however we would also want to include word level n-gram and skipgram features to generate a holistic representation. This establishes a tradeoff between choosing word and character level features. Ideally, one would like to reserve some bits for word and some bits for character level features. We leave the design of the right bit division to future work. ### Effect of Perturbation on Classification
We evaluate LSH projections with text transformations to test whether the projections are robust to input perturbations by nature. We use the character level operations from Section SECREF4. ### Effect of Perturbation on Classification ::: Evaluation Setup
For evaluation, we used the widely popular dialog act and intent prediction datasets. MRDA BIBREF12 is a dialog corpus of multi-party meetings with 6 classes, 78K training and 15K test data; ATIS BIBREF13 is intent prediction dataset for flight reservations with 21 classes, 4.4K training and 893 test examples; and SWDA BIBREF14, BIBREF15 is an open domain dialog corpus between two speakers with 42 classes, 193K training and 5K test examples. For fair comparison, we train LSTM baseline with sub-words and 240 vocabulary size on MRDA, ATIS and SWDA. We uniformly randomly initialized the input word embeddings. We also trained the on-device SGNN model BIBREF6. Then, we created test sets with varying levels of perturbation operations - $\lbrace 20\%,40\%,60\%\rbrace $. ### Effect of Perturbation on Classification ::: Results
Table TABREF15 shows the accuracy results of LSTM and on-device SGNN models. Overall, SGNN models are consistently more robust to perturbations across all three datasets and tasks. One of the reasons is that SGNN relies on word and character level n-gram features, while for LSTMs, the character perturbations result in sub-words being mapped to unknown embedding. This leads LSTM to learn to map inputs with many unknown words to the majority class. We observed the same when we perturbed $100\%$ of the words in the input. As shown in Table TABREF18, the standard deviations of the accuracy with LSTMs are much higher compared to SGNN. This further reinforces the fact that SGNNs are fundamentally more robust to both word misspellings and black box attacks. In the future, we are plan to benchmark SGNN with more aggressive and exploitative black box based attacks. ### Conclusion
In this work, we perform a detailed study analyzing why recent LSH-based projection neural networks are effective for language classification tasks. Through extensive analyses including perturbation studies and experiments on multiple tasks, we show that projection-based neural models are resistant to text transformations compared to widely-used approaches like LSTMs with embeddings. Figure 1: Binary Locality-Sensitive Hashing (LSH) projection representation for text used in SGNN models (Ravi and Kozareva, 2018) Table 1: Collision results: Word level projections Table 2: Collision results: Sentence level projections Table 3: Perturbations with Sentence Projections Table 4: Char. Perturbation with Word Projections Table 5: Comparison of SGNN vs LSTM (using sub-words) after character level perturbations. Table 6: MRDA Accuracy Std. Dev with perturbations | same sentences after applying character level perturbations |
Why is Henry referred to as the First One
A. He was the first man to make it back from a Mars mission.
B. Was the first person pieced back together after death.
C. He was the first American to walk on the surface of Mars.
D. He was the first one to make it back alive from the type of trip that he went on.
| THE FIRST ONE By HERBERT D. KASTLE Illustrated by von Dongen [Transcriber's Note: This etext was produced from Analog July 1961. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed.] The first man to return from beyond the Great Frontier may be welcomed ... but will it be as a curiosity, rather than as a hero...? There was the usual welcoming crowd for a celebrity, and the usual speeches by the usual politicians who met him at the airport which had once been twenty miles outside of Croton, but which the growing city had since engulfed and placed well within its boundaries. But everything wasn't usual. The crowd was quiet, and the mayor didn't seem quite as at-ease as he'd been on his last big welcoming—for Corporal Berringer, one of the crew of the spaceship Washington , first to set Americans upon Mars. His Honor's handclasp was somewhat moist and cold. His Honor's eyes held a trace of remoteness. Still, he was the honored home-comer, the successful returnee, the hometown boy who had made good in a big way, and they took the triumphal tour up Main Street to the new square and the grandstand. There he sat between the mayor and a nervous young coed chosen as homecoming queen, and looked out at the police and fire department bands, the National Guard, the boy scouts and girl scouts, the Elks and Masons. Several of the churches in town had shown indecision as to how to instruct their parishioners to treat him. But they had all come around. The tremendous national interest, the fact that he was the First One, had made them come around. It was obvious by now that they would have to adjust as they'd adjusted to all the other firsts taking place in these—as the newspapers had dubbed the start of the Twenty-first Century—the Galloping Twenties. He was glad when the official greeting was over. He was a very tired man and he had come farther, traveled longer and over darker country, than any man who'd ever lived before. He wanted a meal at his own table, a kiss from his wife, a word from his son, and later to see some old friends and a relative or two. He didn't want to talk about the journey. He wanted to forget the immediacy, the urgency, the terror; then perhaps he would talk. Or would he? For he had very little to tell. He had traveled and he had returned and his voyage was very much like the voyages of the great mariners, from Columbus onward—long, dull periods of time passing, passing, and then the arrival. The house had changed. He saw that as soon as the official car let him off at 45 Roosevelt Street. The change was, he knew, for the better. They had put a porch in front. They had rehabilitated, spruced up, almost rebuilt the entire outside and grounds. But he was sorry. He had wanted it to be as before. The head of the American Legion and the chief of police, who had escorted him on this trip from the square, didn't ask to go in with him. He was glad. He'd had enough of strangers. Not that he was through with strangers. There were dozens of them up and down the street, standing beside parked cars, looking at him. But when he looked back at them, their eyes dropped, they turned away, they began moving off. He was still too much the First One to have his gaze met. He walked up what had once been a concrete path and was now an ornate flagstone path. He climbed the new porch and raised the ornamental knocker on the new door and heard the soft music sound within. He was surprised that he'd had to do this. He'd thought Edith would be watching at a window. And perhaps she had been watching ... but she hadn't opened the door. The door opened; he looked at her. It hadn't been too long and she hadn't changed at all. She was still the small, slender girl he'd loved in high school, the small, slender woman he'd married twelve years ago. Ralphie was with her. They held onto each other as if seeking mutual support, the thirty-three-year old woman and ten-year-old boy. They looked at him, and then both moved forward, still together. He said, "It's good to be home!" Edith nodded and, still holding to Ralphie with one hand, put the other arm around him. He kissed her—her neck, her cheek—and all the old jokes came to mind, the jokes of travel-weary, battle-weary men, the and- then -I'll-put-my-pack-aside jokes that spoke of terrible hunger. She was trembling, and even as her lips came up to touch his he felt the difference, and because of this difference he turned with urgency to Ralphie and picked him up and hugged him and said, because he could think of nothing else to say, "What a big fella, what a big fella." Ralphie stood in his arms as if his feet were still planted on the floor, and he didn't look at his father but somewhere beyond him. "I didn't grow much while you were gone, Dad, Mom says I don't eat enough." So he put him down and told himself that it would all change, that everything would loosen up just as his commanding officer, General Carlisle, had said it would early this morning before he left Washington. "Give it some time," Carlisle had said. "You need the time; they need the time. And for the love of heaven, don't be sensitive." Edith was leading him into the living room, her hand lying still in his, a cool, dead bird lying still in his. He sat down on the couch, she sat down beside him—but she had hesitated. He wasn't being sensitive; she had hesitated. His wife had hesitated before sitting down beside him. Carlisle had said his position was analogous to Columbus', to Vasco De Gama's, to Preshoff's when the Russian returned from the Moon—but more so. Carlisle had said lots of things, but even Carlisle who had worked with him all the way, who had engineered the entire fantastic journey—even Carlisle the Nobel prize winner, the multi-degreed genius in uniform, had not actually spoken to him as one man to another. The eyes. It always showed in their eyes. He looked across the room at Ralphie, standing in the doorway, a boy already tall, already widening in the shoulders, already large of feature. It was like looking into the mirror and seeing himself twenty-five years ago. But Ralphie's face was drawn, was worried in a way that few ten-year-old faces are. "How's it going in school?" he asked. "Gee, Dad, it's the second month of summer vacation." "Well, then, before summer vacation?" "Pretty good." Edith said, "He made top forum the six-month period before vacation, and he made top forum the six-month period you went away, Hank." He nodded, remembering that, remembering everything, remembering the warmth of her farewell, the warmth of Ralphie's farewell, their tears as he left for the experimental flight station in the Aleutians. They had feared for him, having read of the many launchings gone wrong even in continent-to-continent experimental flight. They had been right to worry. He had suffered much after that blow-up. But now they should be rejoicing, because he had survived and made the long journey. Ralphie suddenly said, "I got to go, Dad. I promised Walt and the others I'd pitch. It's Inter-Town Little League, you know. It's Harmon, you know. I got to keep my word." Without waiting for an answer, he waved his hand—it shook; a ten-year-old boy's hand that shook—and ran from the room and from the house. He and Edith sat beside each other, and he wanted badly to take her in his arms, and yet he didn't want to oppress her. He stood up. "I'm very tired. I'd like to lie down a while." Which wasn't true, because he'd been lying down all the months of the way back. She said, "Of course. How stupid of me, expecting you to sit around and make small talk and pick up just where you left off." He nodded. But that was exactly what he wanted to do—make small talk and pick up just where he'd left off. But they didn't expect it of him; they wouldn't let him; they felt he had changed too much. She led him upstairs and along the foyer past Ralphie's room and past the small guest room to their bedroom. This, too, had changed. It was newly painted and it had new furniture. He saw twin beds separated by an ornate little table with an ornate little lamp, and this looked more ominous a barrier to him than the twelve-foot concrete-and-barbed-wire fence around the experimental station. "Which one is mine," he asked, and tried to smile. She also tried to smile. "The one near the window. You always liked the fresh air, the sunshine in the morning. You always said it helped you to get up on time when you were stationed at the base outside of town. You always said it reminded you—being able to see the sky—that you were going to go up in it, and that you were going to come down from it to this bed again." "Not this bed," he murmured, and was a little sorry afterward. "No, not this bed," she said quickly. "Your lodge donated the bedroom set and I really didn't know—" She waved her hand, her face white. He was sure then that she had known, and that the beds and the barrier between them were her own choice, if only an unconscious choice. He went to the bed near the window, stripped off his Air Force blue jacket, began to take off his shirt, but then remembered that some arm scars still showed. He waited for her to leave the room. She said, "Well then, rest up, dear," and went out. He took off his shirt and saw himself in the mirror on the opposite wall; and then took off his under-shirt. The body scars were faint, the scars running in long lines, one dissecting his chest, the other slicing diagonally across his upper abdomen to disappear under his trousers. There were several more on his back, and one on his right thigh. They'd been treated properly and would soon disappear. But she had never seen them. Perhaps she never would. Perhaps pajamas and robes and dark rooms would keep them from her until they were gone. Which was not what he'd considered at all important on leaving Walter Reed Hospital early this morning; which was something he found distasteful, something he felt beneath them both. And, at the same time, he began to understand that there would be many things, previously beneath them both, which would have to be considered. She had changed; Ralphie had changed; all the people he knew had probably changed—because they thought he had changed. He was tired of thinking. He lay down and closed his eyes. He let himself taste bitterness, unhappiness, a loneliness he had never known before. But sometime later, as he was dozing off, a sense of reassurance began filtering into his mind. After all, he was still Henry Devers, the same man who had left home eleven months ago, with a love for family and friends which was, if anything, stronger than before. Once he could communicate this, the strangeness would disappear and the First One would again become good old Hank. It was little enough to ask for—a return to old values, old relationships, the normalcies of the backwash instead of the freneticisms of the lime-light. It would certainly be granted to him. He slept. Dinner was at seven p.m. His mother came; his Uncle Joe and Aunt Lucille came. Together with Edith, Ralphie and himself, they made six, and ate in the dining room at the big table. Before he'd become the First One, it would have been a noisy affair. His family had never been noted for a lack of ebullience, a lack of talkativeness, and Ralphie had always chosen mealtimes—especially with company present—to describe everything and anything that had happened to him during the day. And Edith herself had always chatted, especially with his mother, though they didn't agree about much. Still, it had been good-natured; the general tone of their lives had been good-natured. This wasn't good-natured. Exactly what it was he wasn't sure. "Stiff" was perhaps the word. They began with grapefruit, Edith and Mother serving quickly, efficiently from the kitchen, then sitting down at the table. He looked at Mother as he raised his first spoonful of chilled fruit, and said, "Younger than ever." It was nothing new; he'd said it many many times before, but his mother had always reacted with a bright smile and a quip something like, "Young for the Golden Age Center, you mean." This time she burst into tears. It shocked him. But what shocked him even more was the fact that no one looked up, commented, made any attempt to comfort her; no one indicated in any way that a woman was sobbing at the table. He was sitting directly across from Mother, and reached out and touched her left hand which lay limply beside the silverware. She didn't move it—she hadn't touched him once beyond that first, quick, strangely-cool embrace at the door—then a few seconds later she withdrew it and let it drop out of sight. So there he was, Henry Devers, at home with the family. So there he was, the hero returned, waiting to be treated as a human being. The grapefruit shells were cleaned away and the soup served. Uncle Joe began to talk. "The greatest little development of circular uniform houses you ever did see," he boomed in his powerful salesman's voice. "Still going like sixty. We'll sell out before—" At that point he looked at Hank, and Hank nodded encouragement, desperately interested in this normalcy, and Joe's voice died away. He looked down at his plate, mumbled, "Soup's getting cold," and began to eat. His hand shook a little; his ruddy face was not quite as ruddy as Hank remembered it. Aunt Lucille made a few quavering statements about the Ladies' Tuesday Garden Club, and Hank looked across the table to where she sat between Joe and Mother—his wife and son bracketed him, and yet he felt alone—and said, "I've missed fooling around with the lawn and the rose bushes. Here it is August and I haven't had my hand to a mower or trowel." Aunt Lucille smiled, if you could call it that—a pitiful twitching of the lips—and nodded. She threw her eyes in his direction, and past him, and then down to her plate. Mother, who was still sniffling, said, "I have a dismal headache. I'm going to lie down in the guest room a while." She touched his shoulder in passing—his affectionate, effusive mother who would kiss stray dogs and strange children, who had often irritated him with an excess of physical and verbal caresses—she barely touched his shoulder and fled. So now five of them sat at the table. The meat was served—thin, rare slices of beef, the pink blood-juice oozing warmly from the center. He cut into it and raised a forkful to his mouth, then glanced at Ralphie and said, "Looks fresh enough to have been killed in the back yard." Ralphie said, "Yeah, Dad." Aunt Lucille put down her knife and fork and murmured something to her husband. Joe cleared his throat and said Lucille was rapidly becoming a vegetarian and he guessed she was going into the living room for a while. "She'll be back for dessert, of course," he said, his laugh sounding forced. Hank looked at Edith; Edith was busy with her plate. Hank looked at Ralphie; Ralphie was busy with his plate. Hank looked at Joe; Joe was chewing, gazing out over their heads to the kitchen. Hank looked at Lucille; she was disappearing into the living room. He brought his fist down on the table. The settings jumped; a glass overturned, spilling water. He brought it down again and again. They were all standing now. He sat there and pounded the table with his big right fist—Henry Devers, who would never have thought of making such a scene before, but who was now so sick and tired of being treated as the First One, of being stood back from, looked at in awe of, felt in fear of, that he could have smashed more than a table. Edith said, "Hank!" He said, voice hoarse, "Shut up. Go away. Let me eat alone. I'm sick of the lot of you." Mother and Joe returned a few minutes later where he sat forcing food down his throat. Mother said, "Henry dear—" He didn't answer. She began to cry, and he was glad she left the house then. He had never said anything really bad to his mother. He was afraid this would have been the time. Joe merely cleared his throat and mumbled something about getting together again soon and "drop out and see the new development" and he, too, was gone. Lucille never did manage to speak to him. He finished his beef and waited. Soon Edith came in with the special dessert she'd been preparing half the day—a magnificent English trifle. She served him, and spooned out a portion for herself and Ralphie. She hesitated near his chair, and when he made no comment she called the boy. Then the three of them were sitting, facing the empty side of the table. They ate the trifle. Ralphie finished first and got up and said, "Hey, I promised—" "You promised the boys you'd play baseball or football or handball or something; anything to get away from your father." Ralphie's head dropped and he muttered, "Aw, no, Dad." Edith said, "He'll stay home, Hank. We'll spend an evening together—talking, watching TV, playing Monopoly." Ralphie said, "Gee, sure, Dad, if you want to." Hank stood up. "The question is not whether I want to. You both know I want to. The question is whether you want to." They answered together that of course they wanted to. But their eyes—his wife's and son's eyes—could not meet his, and so he said he was going to his room because he was, after all, very tired and would in all probability continue to be very tired for a long, long time and that they shouldn't count on him for normal social life. He fell asleep quickly, lying there in his clothes. But he didn't sleep long. Edith shook him and he opened his eyes to a lighted room. "Phil and Rhona are here." He blinked at her. She smiled, and it seemed her old smile. "They're so anxious to see you, Hank. I could barely keep Phil from coming up and waking you himself. They want to go out and do the town. Please, Hank, say you will." He sat up. "Phil," he muttered. "Phil and Rhona." They'd had wonderful times together, from grammar school on. Phil and Rhona, their oldest and closest friends. Perhaps this would begin his real homecoming. Do the town? They'd paint it and then tear it down! It didn't turn out that way. He was disappointed; but then again, he'd also expected it. This entire first day at home had conditioned him to expect nothing good. They went to the bowling alleys, and Phil sounded very much the way he always had—soft spoken and full of laughter and full of jokes. He patted Edith on the head the way he always had, and clapped Hank on the shoulder (but not the way he always had—so much more gently, almost remotely), and insisted they all drink more than was good for them as he always had. And for once, Hank was ready to go along on the drinking. For once, he matched Phil shot for shot, beer for beer. They didn't bowl very long. At ten o'clock they crossed the road to Manfred's Tavern, where Phil and the girls ordered sandwiches and coffee and Hank went right on drinking. Edith said something to him, but he merely smiled and waved his hand and gulped another ounce of nirvana. There was dancing to a juke box in Manfred's Tavern. He'd been there many times before, and he was sure several of the couples recognized him. But except for a few abortive glances in his direction, it was as if he were a stranger in a city halfway around the world. At midnight, he was still drinking. The others wanted to leave, but he said, "I haven't danced with my girl Rhona." His tongue was thick, his mind was blurred, and yet he could read the strange expression on her face—pretty Rhona, who'd always flirted with him, who'd made a ritual of flirting with him. Pretty Rhona, who now looked as if she were going to be sick. "So let's rock," he said and stood up. They were on the dance floor. He held her close, and hummed and chatted. And through the alcoholic haze saw she was a stiff-smiled, stiff-bodied, mechanical dancing doll. The number finished; they walked back to the booth. Phil said, "Beddy-bye time." Hank said, "First one dance with my loving wife." He and Edith danced. He didn't hold her close as he had Rhona. He waited for her to come close on her own, and she did, and yet she didn't. Because while she put herself against him, there was something in her face—no, in her eyes; it always showed in the eyes—that made him know she was trying to be the old Edith and not succeeding. This time when the music ended, he was ready to go home. They rode back to town along Route Nine, he and Edith in the rear of Phil's car, Rhona driving because Phil had drunk just a little too much, Phil singing and telling an occasional bad joke, and somehow not his old self. No one was his old self. No one would ever be his old self with the First One. They turned left, to take the short cut along Hallowed Hill Road, and Phil finished a story about a Martian and a Hollywood sex queen and looked at his wife and then past her at the long, cast-iron fence paralleling the road. "Hey," he said, pointing, "do you know why that's the most popular place on earth?" Rhona glanced to the left, and so did Hank and Edith. Rhona made a little sound, and Edith seemed to stop breathing, but Phil went on a while longer, not yet aware of his supposed faux pas . "You know why?" he repeated, turning to the back seat, the laughter rumbling up from his chest. "You know why, folks?" Rhona said, "Did you notice Carl Braken and his wife at—" Hank said, "No, Phil, why is it the most popular place on earth?" Phil said, "Because people are—" And then he caught himself and waved his hand and muttered, "I forgot the punch line." "Because people are dying to get in," Hank said, and looked through the window, past the iron fence, into the large cemetery at the fleeting tombstones. The car was filled with horrified silence when there should have been nothing but laughter, or irritation at a too-old joke. "Maybe you should let me out right here," Hank said. "I'm home—or that's what everyone seems to think. Maybe I should lie down in an open grave. Maybe that would satisfy people. Maybe that's the only way to act, like Dracula or another monster from the movies." Edith said, "Oh, Hank, don't, don't!" The car raced along the road, crossed a macadam highway, went four blocks and pulled to a stop. He didn't bother saying good night. He didn't wait for Edith. He just got out and walked up the flagstone path and entered the house. "Hank," Edith whispered from the guest room doorway, "I'm so sorry—" "There's nothing to be sorry about. It's just a matter of time. It'll all work out in time." "Yes," she said quickly, "that's it. I need a little time. We all need a little time. Because it's so strange, Hank. Because it's so frightening. I should have told you that the moment you walked in. I think I've hurt you terribly, we've all hurt you terribly, by trying to hide that we're frightened." "I'm going to stay in the guest room," he said, "for as long as necessary. For good if need be." "How could it be for good? How, Hank?" That question was perhaps the first firm basis for hope he'd had since returning. And there was something else; what Carlisle had told him, even as Carlisle himself had reacted as all men did. "There are others coming, Edith. Eight that I know of in the tanks right now. My superior, Captain Davidson, who died at the same moment I did—seven months ago next Wednesday—he's going to be next. He was smashed up worse than I was, so it took a little longer, but he's almost ready. And there'll be many more, Edith. The government is going to save all they possibly can from now on. Every time a young and healthy man loses his life by accident, by violence, and his body can be recovered, he'll go into the tanks and they'll start the regenerative brain and organ process—the process that made it all possible. So people have to get used to us. And the old stories, the old terrors, the ugly old superstitions have to die, because in time each place will have some of us; because in time it'll be an ordinary thing." Edith said, "Yes, and I'm so grateful that you're here, Hank. Please believe that. Please be patient with me and Ralphie and—" She paused. "There's one question." He knew what the question was. It had been the first asked him by everyone from the president of the United States on down. "I saw nothing," he said. "It was as if I slept those six and a half months—slept without dreaming." She came to him and touched his face with her lips, and he was satisfied. Later, half asleep, he heard a dog howling, and remembered stories of how they announced death and the presence of monsters. He shivered and pulled the covers closer to him and luxuriated in being safe in his own home. THE END | B. Was the first person pieced back together after death. |
During which period did the Paul Doe undergo radiotherapy?
Choose the correct answer from the following options:
A. 07/14/2014 - 09/06/2014
B. 05/14/2014 - 07/06/2014
C. 07/27/2014 - 09/06/2014
D. 05/14/2014 - 06/30/2014
E. 07/27/2014 - 08/30/2014
| ### Patient Report 0
**Dear colleague, ****Dear colleague, **
We are writing to provide an update regarding Mr. Paul Doe, born on
08/08/1965, who was treated in our clinic from 05/28/14 to 06/20/14.
**Diagnoses: **
- pT1, pN0 (0/21, ECE negative), cM0, Pn0, G2, RX, L0, V0, left
midline tongue carcinoma
- Arterial hypertension
- Status post liver surgery 2013
- Status post endoprosthetic hip treatment
- Idiopathic thrombocytopenia
- Non-insulin-dependent diabetes mellitus type II
- Hypothyroidism
- Nicotine abuse
<!-- -->
- Panendoscopy with sampling on 04/14/2014 and 04/26/2014
**Current Presentation**: With histologically confirmed carcinoma in the
region of the base of the tongue on the left side, Mr. Doe presents for
surgical treatment of the findings. In accordance with the tumor board
decision, resection is performed via a lateral pharyngotomy and neck
dissection on both sides.
**Physical Examination:** Patient in stable general condition (85 kg,
188 cm). MUST score: 0, pain NRS 8/10 intermittent (adjusted with
Acetaminophen) \| fatigue I°, dysphagia I° \| aspiration 0°, ulcer 0°,
trismus 0°, taste disturbance I°, xerostomia I°, osteonecrosis 0°,
hypothyroidism I° (L-thyroxine increased to 150 μg 1-0-0), hoarseness
0°, hearing loss 0° (subjectively reduced), dyspnea: 0°, pneumonitis 0°,
nausea/vomiting 0°.
No suspicious lymph nodes palpable \| movement restrictions 0°,
subcutaneous fibrosis: I°, hyperpigmentation: I° cervical, mucositis 0°,
lymphedema I° (lymphatic drainage prescribed), telangiectasia 0°.
A tumorous mass can be inspected at the base of the left tongue. Tongue
mobility is unremarkable.
**CT chest, abdomen, pelvis on 05/28/14:**
Emphasized mediastinal as well as abdominal lymph
nodes.** **Vasosclerosis. Otherwise, there is no evidence for the
presence of distant metastases with a suspected base of tongue
carcinoma. Liver cirrhosis.
**CT neck on 06/11/14:**
Suspected left tongue base carcinoma crossing midline with extension
into the left vallecula and compression of the left piriform sinus with
suspected lymph node metastases in levels I-III ipsilateral.
Contralateral prominent but not certainly suspicious lymph nodes. The
prominent structure on the left supraclavicular side can also be
interpreted as circumscribed cystiform ectasia of the thoracic duct.
**Ultrasound abdomen on 06/15/14:**
Image of liver cirrhosis status post cholecystectomy.
Hepatosplenomegaly. Moderate aortic sclerosis.
**X-ray pap swallow on 06/17/14:**
Clear tracheal aspiration in the absence of epiglottis envelope. The
cough reflex is preserved. Otherwise, essentially unremarkable
swallowing act.
**Histology**: Invasive, moderately differentiated, squamous cell
carcinoma with keratinization of the medial left base of the tongue,
maximum extent 1.0 cm. Carcinoma- and dysplasia-free biopsies of the
left tonsil, the oropharyngeal tumor/tongue base on the left, deep
resection of the tumor, lower tonsillar pole transition on the left
tongue base, tongue base on the left and medial left tongue base, as
well as median tongue base.
Metastasis-free lymph nodes in Neck-dissection Level IIa to IV on the
left (0/16), Neck-dissection Level IIb on the left (0/1), and
Neck-dissection Level II to IV on the right (0/3), occasionally with
lymphofollicular hyperplasia. Carcinoma-free bone of the left lateral
thigh of the hyoid.
**Final UICC classification:** pT1. pN0 (0/21). L0. V0. Pn0. G2. RX.
**Therapy and Progression**: After the usual clinical and laboratory
preparations, we performed the above-mentioned therapy on 05/29/14 in
intubation anesthesia without complications. For perioperative infection
prophylaxis, the patient received intravenous antibiotic therapy with
Ampicillin and Sulbactam 3g three times daily for the duration of his
hospital stay.
During this procedure, the left lingual artery was interrupted
prophylactically. No postoperative bleeding and no wound healing
disturbances occurred.
A porridge swallow examination showed no evidence of a fistula. On the
following day, the patient was decannulated in consultation with the
colleagues of the speech therapy. After this, a food build-up was
carried out in cooperation with speech therapists. At the time of
discharge, the patient was receiving regular oral nutrition. The stoma
continued to shrink. The patient was monitored, and if necessary, the
tracheostoma was closed with local anesthesia. Histological findings
were pT1. Due to an RX status, adjuvant radiotherapy will be performed
as decided by the tumor board. A prophylactic presentation at the
colleagues of the MKG as a preparatory measure for the upcoming
radiotherapy. We asked for a control re-presentation in our outpatient
clinic on 06/26/14 at 3:00 PM. Further controls take place at the half
and at the end of the radiotherapy and further in 4-6 weeks rhythm. In
case of acute complaints, an immediate re-presentation is possible at
any time.
**Type of surgery**: Lateral pharyngotomy with resection of the base of
the tongue on the left as well as selective neck dissection on both
sides level II-IV with ligature of the lingual artery on the left side,
creation of a stable tracheostoma and tonsillectomy on the left side.
**Surgery report: **First, tracheotomy in a typical manner. A horizontal
incision was made on the skin, positioned approximately two transverse
finger widths above the jugulum. Subsequently, the subcutaneous tissue
and the platysma colli were incised. To facilitate access to the
trachea, the laryngeal muscles were carefully displaced to the side. The
thyroid isthmus was undermined and clamped bilaterally. A precise
transection of the thyroid isthmus followed, with both halves of the
thyroid gland being meticulously sutured using 0- Vicryl. The thyroid
halves were repositioned to expose the trachea. A visceral tracheotomy
was performed, and re-intubation was achieved utilizing a U-tube. The
surgical procedure then transitioned to a neck dissection on the left
side. This phase began with an incision along the anterior edge of the
sternocleidomastoid muscle. The subcutaneous tissue and platysma colli
were carefully cut, with due respect to the auricularis magnus nerve.
Dissection continued dorsally along the sternocleidomastoid muscle to
reach the anterior border of the trapezius muscle. Further exposure
involved the accessorius nerve in a cranialward direction, with
preservation of this neural structure. Dissection proceeded along the
cervical vascular sheath, revealing the common carotid artery, internal
jugular vein, and vagus nerve up to the digastric muscle. Below this
level, exposure of the hypoglossal nerve was achieved.
Successive dissection involved the lymph node fat package, progressing
from level II to level IV in a cranial to caudal and ventral to dorsal
direction. Throughout this process, careful attention was paid to
sparing the aforementioned neural and vascular structures. Subsequently,
access to the lateral pharyngectomy area was gained, allowing
visualization of the external carotid artery along with its branches,
including the superior thyroid artery, superior laryngeal artery, and
lingual artery. Notably, the lingual artery was interrupted during this
stage.
Further exploration revealed the superior laryngeal nerve and
hypoglossal nerve intersecting in a loop above the internal carotid
artery and externally below the external jugular vein. Additional
dissection in a ventral direction followed. The hypoglossal nerve was
prepared meticulously. Exposure of the hyoid bone was achieved, with a
posterior resection of half of the hyoid bone. Importantly, the
hypoglossal nerve was spared during this procedure. Subsequent to these
steps, the lateral pharynx wall was opened, exposing the base of the
hyoid. The next phase of the procedure involved enoral tumor
tonsillectomy on the left side. Starting from the left side, the
surgical team identified the tonsil capsule at the anterior palatal arch
using a Henke spatula. The upper tonsillar pole was then dislodged and
dissected with the Rosenblatt instrument, proceeding from cranial to
caudal. Hemostasis was meticulously achieved through swab pressure and
electrocautery. The excised tonsil tissue was sent for frozen section
examination for further analysis.
**Frozen Section Report: **No evidence of malignancy was found. The
resection was carried out at the junction of the caudal tonsillar pole
and the base of the tongue. At this location, tissue from the base of
the tongue was resected and sent for a frozen section examination, which
revealed no indication of malignancy. Subsequently, a medial resection
of the base of the tongue was performed, confirming the presence of
squamous cell carcinoma in the frozen section analysis. Mucosal suturing
with inverting sutures was then conducted. On the left side, a neck
dissection procedure was performed. The dissection extended along the
sternocleidomastoid muscle, reaching dorsally to the anterior border of
the trapezius muscle. This approach allowed for cranial exposure of the
accessorius nerve while sparing the same. Dissection continued along the
cervical vascular sheath, exposing the common carotid artery, internal
jugular vein, and the vagus nerve up to the digastric muscle. Below
this, the hypoglossal nerve was exposed. Subsequently, the lymph node
fat package was dissected systematically from level II to level IV,
progressing from cranial to caudal and ventral to dorsal, while
carefully preserving the mentioned structures. The surgical procedure
concluded with the placement of a drain, subcutaneous suturing, and skin
suturing.
**Frozen section report:** Invasive squamous cell carcinoma.
**Microscopy:**
Even after paraffin embedding, mucosal cross-sections show a covering of
stratified, non-keratinizing squamous epithelium with occasional
significant stratification disturbances extending into superficial cell
layers. This transitions into invasive growth with solid clusters of
polygonal tumor cells, some of which exhibit identifiable intercellular
bridges. The cell nuclei are enlarged, round to oval, with occasional
small nucleoli and mild to moderate nuclear pleomorphism. Dyskeratosis
is observed in some areas.
**Lab results upon Discharge: **
**Parameter** **Result** **Reference Range**
-------------------- ----------------------- -----------------------
Sodium 141 mEq/L 135 - 145 mEq/L
Potassium 4.7 mEq/L 3.5 - 5.0 mEq/L
Creatinine 1.0 mg/dL 0.7 - 1.3 mg/dL
Calcium 9.04 mg/dL 8.8 - 10.6 mg/dL
GFR (MDRD) \> 60 mL/min/1.73m\^2 \> 60 mL/min/1.73m\^2
GFR (CKD-EPI,CREA) 80 mL/min/1.73m\^2 \> 90 mL/min/1.73m\^2
C-reactive protein 1.0 mg/dL \< 0.5 mg/dL
### Patient Report 1
**Dear colleague, **
We are writing to provide an update regarding Mr. Paul Doe, born on
08/08/1965, who presented to our outpatient clinic on 09/14/2014.
**Diagnosis: **Tongue base Carcinoma ICD-10: C01, stage: pT1 pN0 (0/21)
L0 V0 Pn0 G2 RX
**Other Diagnoses:**
- Arterial hypertension
- Diabetes mellitus type II
- Status post liver surgery 2013
- Status post endoprosthetic hip treatment
- Hypothyroidism
- Oral thrush
- Hypacusis
- Since 03/2014: Odynophagia
<!-- -->
- 05/14/2014: Panendoscopy, biopsy, and initial diagnosis of left
tongue base carcinoma.
- 05/29/2014: Tumor resection and selective neck dissection LI-III
**Histology:**
Invasive, moderately differentiated, squamous cell carcinoma with
keratinization of the medial left base of the tongue, maximum extent 1.0
cm. Carcinoma- and dysplasia-free biopsies of the left tonsil, the
oropharyngeal tumor/tongue base on the left, deep resection of the
tumor, lower tonsillar pole transition on the left tongue base, tongue
base on the left and medial left tongue base, as well as median tongue
base.
Metastasis-free lymph nodes in Neck-dissection Level IIa to IV on the
left (0/16), Neck-dissection Level IIb on the left (0/1), and
Neck-dissection Level II to IV on the right (0/3), occasionally with
lymphofollicular hyperplasia. Carcinoma-free bone of the left lateral
thigh of the hyoid.
**Final UICC classification:** min pT1, pN0 (0/21). L0. V0. Pn0. G2. RX.
**Current Radiotherapy:**
**Indication**: According to the decision made by the interdisciplinary
tumor board for head and neck tumors, it was determined by our medical
team that, in the postoperative condition following the resection of a
tongue base carcinoma with an unclear resection status, there is an
indication for radiation therapy of the former tumor site.
**Technique:** Percutaneous radiotherapy of the former primary tumor
region with 6-MeVPhotons, in Rapid-Arc technique, with a single dose of
2 Gy up to a total dose of 60 Gy.
**Radiotherapy 07/27/2014 - 09/06/2014:**
During the course of radiotherapy, the patient experienced enoral
mucositis (grade II according to CTCAE) leading to subsequent
odynophagia and dysphagia. We managed these symptoms with oral rinses
and initiated pain management using Acetaminophen, resulting in an
acceptable reduction of pain over time. At the end of the therapy, the
patient\'s general condition remained stable (ECOG performance status:
70%). Second-degree mucositis enoral persisted, causing ongoing
dysphagia and odynophagia. Additionally, the patient exhibited localized
radiodermatitis (grade II according to CTCAE) within the radiation
field. The patient did not report xerostomia or dysgeusia.
**Current Recommendations:**
The patient received comprehensive instructions on continued skincare
and side-effect management. An initial follow-up appointment with the
radio-oncology team has been scheduled in our outpatient clinic. We
kindly request the patient to provide a renewed referral for
radiotherapy on the day of the appointment.
The ongoing oncological treatment plan will be determined by the
patient\'s Ear, Nose, and Throat specialists. Regular follow-up
examinations are strongly recommended. Additionally, for patients who
have completed radiation therapy in the ENT region, we advise lifelong
adherence to fluoride prophylaxis and antibiotic therapy during any
dental procedures.
### Patient Report 2
**Dear colleague, **
We report about our patient, Mr. Doe, born on 08/08/1965, who presented
to our outpatient clinic for phoniatrics and pedaudiology on 10/10/2014.
**Diagnoses: **
- Tongue base carcinoma ICD-10: C01, stage: pT1 pN0 (0/21) L0 V0 Pn0
G2 RX
- Since 03/2014: Odynophagia
- 05/14/2014: Panendoscopy, biopsy, and initial diagnosis of left
tongue base carcinoma.
- 05/29/2014: Tumor resection and selective neck dissection LI-III
**Other Diagnoses:**
- Arterial hypertension
- Diabetes mellitus type II
- Status post liver surgery 2013
- Status post endoprosthetic hip treatment
- Hypothyroidism
- Oral thrush
- Hypacusis
**Medical History: **We may kindly assume the detailed history as known.
**Phoniatrics: Fiberoptic endoscopic swallow examination:**
Tongue motor function is well preserved, sensitivity of lip and tongue
laterally equal,
Mucous membranes non-irritant on all sides. Tracheal mucosa
non-irritant, no evidence of saliva intratracheal. Neopharynx
inconspicuous, air bubbles visible above Provox outlet on pressing
attempt. Tongue retraction slightly limited. Velopharyngeal closure
good.
**Therapy and Course:** After completion of the adjuvant radiotherapy
approximately 6 weeks ago, phonation via the Provox voice prosthesis was
no longer possible after this had initially worked after the operation.
Additionally, there were issues with regurgitation of ingested
substances, regardless of their consistency. In some cases, nasal
penetration with fluids occurred. There were no indications of
aspiration. A self-assessment, involving the use of blue-colored liquid,
revealed no signs of leakage from the Provox device.
**Current Recommendations:** Oncological follow-up in 12 months.
### Patient Report 3
**Dear colleague, **
We report about our patient, Mr. Doe, born on 08/08/1965 who presented
at our outpatient clinic for radio-oncological follow-up on 10/09/2020.
**Diagnoses: **
- Tongue base Carcinoma ICD-10: C01, stage: pT1 pN0 (0/21) L0 V0 Pn0
G2 RX
- Since 03/2014: Odynophagia
- 05/14/2014: Panendoscopy, biopsy, and initial diagnosis of left
tongue base carcinoma.
- 05/29/2014: Tumor resection and selective neck dissection LI-III
**Other Diagnoses:**
- Arterial hypertension
- Diabetes mellitus type II
- Status post liver surgery 2013
- Status post endoprosthetic hip treatment
- Hypothyroidism
- Oral thrush
- Hypacusis
**Current Presentation: **The patient presented to our general
outpatient clinic for a radio-oncological follow-up on 10/09/2020 in the
presence of his wife.
**Physical Examination**: Patient in stable general condition (85 kg,188
cm). MUST score: 0, pain NRS 8/10 intermittent (adjusted with
Acetaminophen) \| fatigue I°, dysphagia I° \| aspiration 0°, ulcer 0°,
trismus 0°, taste disturbance I°, xerostomia I°, osteonecrosis 0°,
hypothyroidism I°, hoarseness 0°, hearing loss 0° (subjectively
reduced), dyspnea: 0°, pneumonitis 0°, nausea/vomiting 0°.
No suspicious lymph nodes palpable \| movement restrictions 0°,
subcutaneous fibrosis: I°, hyperpigmentation: I° cervical, mucositis 0°,
lymphedema I° (lymphatic drainage prescribed), telangiectasia 0°.
**MRI scan of the neck from 10/09/2020:**
Clear post-therapeutic changes in the resection and radiation area after
adjuvant RTx following tumor resection with laryngectomy for extensive
recurrence of oropharyngeal cancer.
Size constant, but still clearly accentuated lymph nodes in level Ib/IIa
on the left. Regredience of seroma formation under the left
sternocleidomastoid muscle.
**Current Recommendations: **Primary oncological care and follow-up,
including imaging, will be provided by the ENT clinic according to the
guidelines. A re-appointment for a further radio-oncological follow-up
at the follow-up appointment at the Radiation Therapy Tumor Therapy
Center has been scheduled. After head and neck radiation therapy,
regular fluoridation of the teeth and guideline-based antibiotic
prophylaxis is required prior to major dental procedures. We also
recommend temporomandibular joint opening exercises to prevent
temporomandibular joint fibrosis and consecutive temporomandibular joint
opening obstruction. We also refer to regular control of thyroid
function parameters and, if necessary, initiation of substitution
therapy after radiotherapy to the neck.
### Patient Report 4
**Dear colleague, **
We hereby report on our patient Mr. Paul Doe, born 08/08/1965 for
radio-oncological follow-up on 09/24/2021.
**Diagnoses: **
- Tongue base carcinoma ICD-10: C01, stage: pT1 pN0 (0/21) L0 V0 Pn0
G2 RX
- Since 03/2014: Odynophagia
- 05/14/2014: Panendoscopy, biopsy, and initial diagnosis of left
tongue base carcinoma.
- 05/29/2014: Tumor resection and selective neck dissection LI-III
**Other Diagnoses:**
- Arterial hypertension
- Diabetes mellitus type II
- Status post liver surgery 2013
- Status post endoprosthetic hip treatment
- Hypothyroidism
- Oral thrush
- Hypacusis
**Current Presentation: **The patient presented to our general
outpatient clinic for radio-oncological follow-up on 09/24/2021.
**Physical Examination: **Patient in reduced general condition (KPS 60%,
86 kg,188 cm). Weight loss 0°, MUST score: 0, pain VAS 1-2/10, fatigue
I°, dysphagia I° with solid food (liquids occasionally flow out of the
nose again when swallowing), aspiration 0°, ulcer 0°, trismus 0°, taste
disorder I° (present in approx. 80%), xerostomia I°, osteonecrosis 0°,
hypothyroidism II°, hoarseness II°, hearing loss, I°, dyspnea: 0°,
pneumonitis 0°, nausea/vomiting 0°.
No suspicious lymph nodes palpable, movement restrictions I° head
reclination restricted with tension and pain, subcutaneous fibrosis: I°,
hyperpigmentation: I°, mucositis 0°, lymphedema I° (lymphatic drainage),
telangiectasia 0°.
**MR neck plain + contrast agent on 09/24/2021**:
[Technique]{.underline}: STIR triplanar, T1 ax -/+ contrast agent, T1
mDixon cor after
contrast agent.
[Findings]{.underline}: Known status post tumor resection with
laryngectomy for extensive recurrence of oropharyngeal Carcinoma;
Follow-up after RTx. Somewhat increasing swelling of nasopharynx to
oropharynx. From the uvula the swelling is stable. As far as can be
assessed, no clear recurrence-specific tissue proliferation or contrast
uptake. Unchanged accentuated lymph nodes in level Ib/IIa on the left
(one exemplary measured lymph node borderline large, idem to preliminary
examination). Mastoid cells minimally displaced on the left. Moderate
degenerative changes of the cervical spine. Assessment. Increasing
swelling of the naso- to oropharynx. Neopharynx unchanged swollen. No
evidence of malignancy-suspicious lymph nodes.
**CT scan of the thorax on 09/24/2021**:
Size-constant visualization of interlobar oval compaction in the left
upper lobe corresponding to an interlobar lymph node. New to the
previous examination, two small nodular condensations appear, basal in
the right and in the left lower lobe, differentially inflammatory;
follow-up is recommended. Unchanged the prominent mediastinal lymph
nodes, constant in size and number.
**Current Recommendations:**
Primary oncologic care and follow-up including imaging will take place
via the ENT clinic on 01/14/22 at 11:00 AM. A re-appointment for the
next radio-oncological follow-up has been arranged for 01/14/2022 at
1:00 PM in our radiotherapy outpatient clinic in the Tumor Therapy
Center.
**Lab results upon Discharge:**
**Hematology**
**Parameter ** **Result** **Reference**
---------------- ------------- -----------------------
WBC 6,900 /μL 4,500 - 11,000 /μL
RBC 2.7M /μL 4.5M - 5.9M /μL
Hemoglobin 8.2 g/dL 14 - 18 g/dL
Hematocrit 25.1 % 40 - 48 %
MCH 31.27 pg 27 - 33 pg
MCV 94 fL 82 - 92 fL
MCHC 32.7 g/dL 32 - 36 g/dL
Platelets 638,000 /μL 150,000 - 450,000 /μL
**Serum chemistry**
**Parameter ** **Result** **Reference**
---------------- ------------ -----------------
Sodium 144 mEq/L 135 - 145 mEq/L
Potassium 4.8 mEq/L 3.5 - 5.0 mEq/L
Creatinine 1.3 mg/dL 0.7 - 1.3 mg/dL
ALT 21 U/L 10 - 50 U/L
eGFR 55 mL/min \> 90 mL/min
CRP 2.9 mg/dL \< 0.5 mg/dL
**Coagulation**
**Parameter ** **Result** **Reference**
---------------- ------------ ---------------
PT 93 % 70 - 120 %
INR 1.1 0.8 - 1.2
aPTT 31 sec 26 - 37 sec
**Thyroid hormones**
**Parameter ** **Result** **Reference**
---------------- ------------- ------------------
TSH 1.16 μIU/mL 0.4 - 4.2 μIU/mL
fT3 2.38 pg/mL 2.3 - 4.2 pg/mL
fT4 1.70 ng/dL 0.9 - 1.7 ng/dL
### Patient Report 5
**Dear colleague, **
We are reporting on Mr. Paul Doe, born on 08/08/1965, who was admitted
to our hospital from 01/10/2022 to 01/27/2022.
**Diagnosis**: Metachronous pulmonary metastatic squamous cell carcinoma
**Diagnoses: **
- Tongue base Carcinoma ICD-10: C01, stage: pT1 pN0 (0/21) L0 V0 Pn0
G2 RX
- Since 03/2014 Odynophagia
- 05/14/2014 Panendoscopy, biopsy, and initial diagnosis of left
tongue base carcinoma.
- 05/29/2014 Tumor resection and selective neck dissection LI-III
**Other Diagnoses:**
- Arterial hypertension
- Diabetes mellitus type II
- Status post liver surgery 2013
- Status post endoprosthetic hip treatment
- Hypothyroidism
- Oral thrush
- Hypacusis
**Planned Surgical Procedure:**
- Perioperative bronchoscopy
- Left-sided video-assisted thoracoscopic surgery
- Pleurolysis
- Anatomical upper lobe resection
- Systematic mediastinal, hilar, and interlobar lymph node dissection
- Placement of double chest drains
**Medical History**: Mr. Paul Doe initially presented with a diagnosis
of pT1, pN0 (0/21, ECE negative), cM0, Pn0, G2, RX, L0, V0, left midline
tongue carcinoma. Panendoscopy with specimen collection on 04/14/2014
and 04/26/2014 confirmed carcinoma.
Surgical resection was performed via lateral pharyngotomy and neck
dissection. Histology confirmed squamous cell carcinoma. He subsequently
received radiotherapy.
During a follow-up CT examination, a suspicious lesion was identified in
the lung, which raised concerns regarding the possibility of metastasis
originating from the previously diagnosed left midline tongue carcinoma
(pT1, pN0, G2, RX).
**Current Presentation:** Mr. Doe was admitted for further examination
and treatment to assess the behavior and extent of the lesion.
Clinically, Mr. Doe was in stable general condition and had no symptoms
suggestive of B symptoms.
**Therapy and Progression**: The above-mentioned procedure was performed
without complications on 01/10/2022. Histologically, the final resection
specimen confirmed the presence of a 2.9 cm squamous cell carcinoma,
consistent with a metastatic recurrence of the previously known
hypopharyngeal carcinoma. The dissected lymph nodes were free of tumor.
The postoperative course was uneventful. In the absence of any
complications, surgical sutures were removed on day 10 after surgery. A
current chest X-ray showed a regular postoperative outcome with
sufficient expansion of the left lobe. Further monitoring is recommended
by the treating colleagues in this regard. Mr. Doe is also connected to
the radiation therapy team for ongoing follow-up.
If there are any complications or questions, please contact the relevant
ward or reach out to our Central Patient Management. Outside regular
working hours, you can contact the on-call colleague in the Abdominal
Surgery department for assistance.
For further information on the patient\'s discharge management, treating
providers are available for inquiries from Monday to Friday, 9 AM to 7
PM, as well as on weekends and holidays from 10 AM to 2 PM.
Mr. Doe was discharged from the hospital on 01/27/2022.
**Addition:**
**Histology Report:** Resected left upper lobe specimen with a 2.9 cm
solid carcinoma. The histological picture is consistent with a
metastasis from the previously diagnosed non-keratinizing squamous cell
carcinoma. There were focal vascular invasions. No pleural invasion was
observed. The resection was complete, with all dissected lymph nodes
showing no tumor involvement.
**Chest X-ray, anterior-posterior view from 01/21/2022**:
[Clinical Information:]{.underline} History of VATS with wedge
resection, yesterday\'s drain removal
[Question:]{.underline} Follow-up, pneumothorax after drain removal?
Infiltrates? Atelectasis?
[Findings:]{.underline} Left chest drainage tube has been removed. Left
apical pneumothorax line, measuring approximately 2.3 cm. Continued
extensive shadowing of the left upper field, most likely postoperative,
infiltrate cannot be definitively ruled out. Slightly hypotransparent
left lung in comparison, most likely due to residual postoperative
reduced ventilation. No effusion. Widened cardiac silhouette. Regression
of dystelectasis in the right lower field. No acute signs of pulmonary
venous congestion. Trachea is mid-positioned and not stenosed.
Left-sided port catheter still in place.
[Summary]{.underline}: Left apical corax with a width of 2.3 cm.
Continued extensive shadowing of the left upper field, most likely
postoperative, with infiltrate not definitively excluded. Residual
postoperative reduced ventilation on the left side. Regression of
dystelectasis in the right lower field. No effusion. No acute signs of
pulmonary venous congestion. Follow-up recommended.
**Examinations Chest X-ray, anterior-posterior view from 01/23/2022:**
[Question]{.underline}**:** Follow-up.
[Findings]{.underline}: Left apical pneumothorax, measuring
approximately 1.4 cm. Extensive shadowing in projection onto the left
upper lobe, differentials include postoperative changes, incipient
infiltrate not excluded. Dystelectasis of the right lower field. No
evidence of pleural effusion or acute pulmonary venous congestion.
Cardiomegaly. Indwelling chest drainage with the catheter tip projecting
onto the left upper lobe. Well-positioned port catheter tip projecting
onto the right atrial entrance plane. No evidence of pleural effusion.
Assessment Left apical pneumothorax, measuring approximately 1.4 cm,
with indwelling left chest drainage. Extensive shadowing in projection
onto the left upper lobe, differentials include postoperative changes,
incipient infiltrate not excluded. Dystelectasis of the right lower
field. No significant pleural effusion. No acute pulmonary venous
congestion. Cardiomegaly.
**Chest X-ray, anterior-posterior view from 01/25/2022**
[Previous Examinations]{.underline}: Appearance of a diffuse
postoperative shadow in the left upper field. No evidence of pleural
effusion, inflammatory infiltrate, or pulmonary venous congestion.
[Findings]{.underline}: Left-sided chest port with the tip projecting
onto the superior vena cava. The upper mediastinum is narrow, the
trachea is mid-positioned and patent.
[Assessment]{.underline}: The pneumothorax appears to be largely
resolved.
**Urinanalysis**:
Material: Urine, midstream sample collected on 01/11/2022
- Antimicrobial inhibitors negative
- No evidence of growth-inhibiting substances in the sample material.
- Colony Count (CFU) / mL \<1,000, Assessment: A low colony count
typically does not support a urinary tract infection.
- Epithelial cells (microscopic) \<20 epithelial cells/μL
- Leukocytes (microscopic) \<20 leukocytes/μL
- Microorganisms (microscopic) 20-100 microorganisms/μL Pathogen
Enterococci
**Lab values upon Discharge: **
**Parameter** **Result** **Reference Range**
-------------------- --------------------- ---------------------
Sodium 141 mEq/L 135 - 145 mEq/L
Potassium 4.7 mEq/L 3.5 - 5.0 mEq/L
Creatinine 1.1 mg/dL 0.7 - 1.3 mg/dL
Calcium 10.4 mg/dL 8.8 - 10.6 mg/dL
eGFR (MDRD) \> 60 mL/min/1.73m² \> 60 mL/min/1.73m²
eGFR (CKD-EPI) 85 mL/min/1.73m² \> 90 mL/min/1.73m²
C-Reactive Protein 5.0 mg/dL \< 0.5 mg/dL
### Patient Report 6
**Dear colleague, **
We are reporting on Mr. Paul Doe, born on 08/08/1965, who presented to
our surgical outpatient clinic on 01/24/2022.
**Diagnoses**: Metachronous pulmonary metastatic squamous cell carcinoma
at the base of the tongue.
**Surgical Procedure from 01/11/2022:**
- Perioperative bronchoscopy
- Left-sided video-assisted thoracoscopic surgery
- Pleurolysis
- Anatomical upper lobe resection
- Systematic mediastinal, hilar, and interlobar lymph node dissection
- Placement of double chest drains
**Previous Diagnoses and Therapies:**
- Recurrent oropharyngeal carcinoma ICD-10: C01
- Stage: rpT2 rpN2(2/24, ECE -) L1 V0 Pn0 G2 R0
- Tumor localization: base of tongue, crossing midline
- Since 04/2014: Odynophagia
- 04/14/2014: Panendoscopy, biopsy and initial diagnosis of left
tongue base carcinoma.
- 05/29/2014: Tumor resection and selective neck dissection LI-III
- 08-10/2015: radiotherapy of former PTR with 60 Gy à 2 Gy.
- OSAS with CPAP incompliance
<!-- -->
- Liver cirrhosis with alcohol abuse
- Non-insulin-dependent diabetes mellitus type II
- Arterial hypertension
- History of endoprosthetic hip treatment
- Hypothyroidism
- Oral thrush
- Hypacusis
**Current Presentation:** Mr. Doe presented for postoperative follow-up
after his left-sided videothoracoscopic upper lobe resection due to
previously diagnosed pulmonary metastatic hypopharyngeal carcinoma.
**Medical History:** The patient\'s general condition is good. He is
currently on intermittent as-needed analgesia with Acetaminophen. The
final resection specimen histologically confirmed the presence of a 2.9
cm squamous cell carcinoma, consistent with a metastatic recurrence of
the previously known hypopharyngeal carcinoma. The dissected lymph nodes
were free of tumor.
**Therapy and Progression**: During the follow-up appointment, Mr. Doe
underwent a clinical examination and a chest X-ray to assess his
postoperative condition. The examination revealed no new concerning
findings, and Mr. Doe continued to remain in stable general condition.
His surgical incision site was inspected, showing signs of satisfactory
healing without any signs of infection or complications.
Furthermore, Mr. Doe\'s lung function was evaluated through spirometry,
which indicated adequate pulmonary function post-surgery. He was also
provided with personalized recommendations for respiratory exercises to
optimize his lung function during the recovery period.
**Current Recommendations:**
- Mr. Doe is connected to the radiation therapy team for ongoing
follow-up.
### Patient Report 7
**Dear colleague, **
We are reporting on Mr. Paul Doe, born on 08/08/1965, who presented to
our surgical outpatient clinic on 01/24/2022.
**Diagnoses**: Metachronous pulmonary metastatic squamous cell carcinoma
at the base of the tongue.
**Surgery on 01/11/2022:**
- Left-sided video-assisted thoracoscopic surgery
- Anatomical upper lobe resection
- Systematic mediastinal, hilar, and interlobar lymph node dissection
**Previous Diagnoses and Therapies:**
- Recurrent oropharyngeal carcinoma ICD-10: C01
- Stage: rpT2 rpN2(2/24, ECE -) L1 V0 Pn0 G2 R0
- Tumor localization: base of tongue, crossing midline
- Since 04/2014: Odynophagia
- 04/14/2014: Panendoscopy, biopsy and initial diagnosis of left
tongue base carcinoma.
- 05/29/2014: Tumor resection and selective neck dissection LI-III
- 08-10/2015 Radiotherapy of former PTR with 60 Gy à 2 Gy.
**Other Diagnoses:**
- OSAS with CPAP incompliance
- Liver cirrhosis with alcohol abuse
- Non-insulin-dependent diabetes mellitus type II
- Arterial hypertension
- History of endoprosthetic hip treatment
- Hypothyroidism
- Oral thrush
- Hypacusis
**Current Presentation:** Mr. Doe presented for routine follow-up.
Clinically, he remains in stable general condition, with no signs of
B-symptoms.
**Medical History**: The surgical procedure performed on 01/11/2022
involved perioperative bronchoscopy, left VATS, pleurolysis, anatomical
upper lobe resection, systematic mediastinal, hilar, and interlobar
lymph node dissection, as well as the placement of double chest drains.
Histologically, the final resection specimen confirmed a 2.9 cm solid
carcinoma, consistent with metastasis from the previously diagnosed
squamous cell carcinoma. Lymph nodes dissected during the procedure were
tumor-free. Mr. Doe\'s postoperative course was uneventful, and surgical
sutures were removed on day 10 after surgery.
**Physical Examination:** Patient in good general condition. Weight loss
0°, MUST score: 0, pain VAS 1-2/10, fatigue I°, dysphagia I° with solid
food (liquids occasionally flow out of the nose again when swallowing),
aspiration 0°, ulcer 0°, trismus 0°, taste disorder I° (present in
approx. 80%), xerostomia I°, osteonecrosis 0°, hypothyroidism II°,
hoarseness II°, hearing loss, I°, dyspnea: 0°, pneumonitis 0°,
nausea/vomiting 0°.
No suspicious lymph nodes palpable, movement restrictions I° head
reclination restricted with tension and pain, subcutaneous fibrosis: I°,
hyperpigmentation: I°, mucositis 0°, lymphedema I° (lymphatic drainage),
telangiectasia 0°.
**Current Recommendations:** Mr. Doe is advised to continue his
follow-up appointments.
**Lab results upon Discharge:**
**Parameter** **Results** **Reference Range**
-------------------------------- -------------- ---------------------
Neutrophils 72.2 % 42.0-77.0 %
Lymphocytes 8.6 % 20.0-44.0 %
Monocytes 11.6 % 2.0-9.5 %
Basophils 1.4 % 0.0-1.8 %
Eosinophils 6.0 % 0.5-5.5 %
Immature Granulocytes 0.2 % 0.0-1.0 %
Sodium 137 mEq/L 136-145 mEq/L
Potassium 4.2 mEq/L 3.5-4.5 mEq/L
Calcium 9.24 mg/dL 8.8-10.2 mg/dL
Chloride 100 mEq/L 98-107 mEq/L
Creatinine 1.27 mg/dL 0.70-1.20 mg/dL
BUN 48 mg/dL 17-48 mg/dL
Uric Acid 5.2 mg/dL 3.6-8.2 mg/dL
CRP 0.8 mg/L \< 5.0 mg/L
PSA 2.31 ng/mL \< 4.40 ng/mL
ALT 12 U/L \< 41 U/L
AST 38 U/L \< 50 U/L
Alkaline Phosphatase 115 U/L 40-130 U/L
GGT 20 U/L 8-61 U/L
LDH 335 U/L 135-250 U/L
Testosterone \<0.03 ng/mL 1.32-8.92 ng/mL
TSH 1.42 mIU/L 0.27-4.20 mIU/L
Hemoglobin 10.1 g/dL 12.5-17.2 g/dL
Hematocrit 28.5 % 37.0-49.0 %
RBC 3.3 M/µL 4.0-5.6 M/µL
WBC 4.98 K/µL 3.90-10.50 K/µL
Platelets 281 K/µL 150-370 K/µL
MCV 85.6 fL 80.0-101.0 fL
MCH 30.3 pg 27.0-34.0 pg
MCHC 35.4 g/dL 31.5-36.0 g/dL
MPV 9.2 fL 7.0-12.0 fL
RDW 13.4 % 11.5-15.0 %
Absolute Neutrophils 3.59 K/µL 1.50-7.70 K/µL
Absolute Immature Granulocytes 0.010 K/µL \< 0.050 K/µL
Absolute Lymphocytes 0.43 K/µL 1.10-4.50 K/µL
Absolute Monocytes 0.58 K/µL 0.10-0.90 K/µL
Absolute Eosinophils 0.30 K/µL 0.02-0.50 K/µL
Absolute Basophils 0.07 K/µL 0.00-0.20 K/µL
Reticulocytes 31.3 K/µL 25.0-105.0 K/µL
Reticulocyte % 0.94 % 0.50-2.00 %
Ret-Hb 33.9 pg 28.5-34.5 pg
PT 112 % \> 78 %
INR 0.95 \< 1.25
aPTT 30.2 sec. 25.0-38.0 sec. | 07/27/2014 - 09/06/2014 |
Why does Paul think an alien wouldn't be able to hide on Earth?
A. Aliens don't look like Earthlings.
B. An alien would not be able to mimic a human enough to fit in with society.
C. An alien wouldn't be able to assimilate into Earth's backward culture.
D. The Earth has so many intelligence agencies, at least one would be watching when an alien gave itself away.
| One can't be too cautious about the people one meets in Tangier. They're all weirdies of one kind or another. Me? Oh, I'm A Stranger Here Myself By MACK REYNOLDS The Place de France is the town's hub. It marks the end of Boulevard Pasteur, the main drag of the westernized part of the city, and the beginning of Rue de la Liberté, which leads down to the Grand Socco and the medina. In a three-minute walk from the Place de France you can go from an ultra-modern, California-like resort to the Baghdad of Harun al-Rashid. It's quite a town, Tangier. King-size sidewalk cafes occupy three of the strategic corners on the Place de France. The Cafe de Paris serves the best draft beer in town, gets all the better custom, and has three shoeshine boys attached to the establishment. You can sit of a sunny morning and read the Paris edition of the New York Herald Tribune while getting your shoes done up like mirrors for thirty Moroccan francs which comes to about five cents at current exchange. You can sit there, after the paper's read, sip your expresso and watch the people go by. Tangier is possibly the most cosmopolitan city in the world. In native costume you'll see Berber and Rif, Arab and Blue Man, and occasionally a Senegalese from further south. In European dress you'll see Japs and Chinese, Hindus and Turks, Levantines and Filipinos, North Americans and South Americans, and, of course, even Europeans—from both sides of the Curtain. In Tangier you'll find some of the world's poorest and some of the richest. The poorest will try to sell you anything from a shoeshine to their not very lily-white bodies, and the richest will avoid your eyes, afraid you might try to sell them something. In spite of recent changes, the town still has its unique qualities. As a result of them the permanent population includes smugglers and black-marketeers, fugitives from justice and international con men, espionage and counter-espionage agents, homosexuals, nymphomaniacs, alcoholics, drug addicts, displaced persons, ex-royalty, and subversives of every flavor. Local law limits the activities of few of these. Like I said, it's quite a town. I looked up from my Herald Tribune and said, "Hello, Paul. Anything new cooking?" He sank into the chair opposite me and looked around for the waiter. The tables were all crowded and since mine was a face he recognized, he assumed he was welcome to intrude. It was more or less standard procedure at the Cafe de Paris. It wasn't a place to go if you wanted to be alone. Paul said, "How are you, Rupert? Haven't seen you for donkey's years." The waiter came along and Paul ordered a glass of beer. Paul was an easy-going, sallow-faced little man. I vaguely remembered somebody saying he was from Liverpool and in exports. "What's in the newspaper?" he said, disinterestedly. "Pogo and Albert are going to fight a duel," I told him, "and Lil Abner is becoming a rock'n'roll singer." He grunted. "Oh," I said, "the intellectual type." I scanned the front page. "The Russkies have put up another manned satellite." "They have, eh? How big?" "Several times bigger than anything we Americans have." The beer came and looked good, so I ordered a glass too. Paul said, "What ever happened to those poxy flying saucers?" "What flying saucers?" A French girl went by with a poodle so finely clipped as to look as though it'd been shaven. The girl was in the latest from Paris. Every pore in place. We both looked after her. "You know, what everybody was seeing a few years ago. It's too bad one of these bloody manned satellites wasn't up then. Maybe they would've seen one." "That's an idea," I said. We didn't say anything else for a while and I began to wonder if I could go back to my paper without rubbing him the wrong way. I didn't know Paul very well, but, for that matter, it's comparatively seldom you ever get to know anybody very well in Tangier. Largely, cards are played close to the chest. My beer came and a plate of tapas for us both. Tapas at the Cafe de Paris are apt to be potato salad, a few anchovies, olives, and possibly some cheese. Free lunch, they used to call it in the States. Just to say something, I said, "Where do you think they came from?" And when he looked blank, I added, "The Flying Saucers." He grinned. "From Mars or Venus, or someplace." "Ummmm," I said. "Too bad none of them ever crashed, or landed on the Yale football field and said Take me to your cheerleader , or something." Paul yawned and said, "That was always the trouble with those crackpot blokes' explanations of them. If they were aliens from space, then why not show themselves?" I ate one of the potato chips. It'd been cooked in rancid olive oil. I said, "Oh, there are various answers to that one. We could probably sit around here and think of two or three that made sense." Paul was mildly interested. "Like what?" "Well, hell, suppose for instance there's this big Galactic League of civilized planets. But it's restricted, see. You're not eligible for membership until you, well, say until you've developed space flight. Then you're invited into the club. Meanwhile, they send secret missions down from time to time to keep an eye on your progress." Paul grinned at me. "I see you read the same poxy stuff I do." A Moorish girl went by dressed in a neatly tailored gray jellaba, European style high-heeled shoes, and a pinkish silk veil so transparent that you could see she wore lipstick. Very provocative, dark eyes can be over a veil. We both looked after her. I said, "Or, here's another one. Suppose you have a very advanced civilization on, say, Mars." "Not Mars. No air, and too bloody dry to support life." "Don't interrupt, please," I said with mock severity. "This is a very old civilization and as the planet began to lose its water and air, it withdrew underground. Uses hydroponics and so forth, husbands its water and air. Isn't that what we'd do, in a few million years, if Earth lost its water and air?" "I suppose so," he said. "Anyway, what about them?" "Well, they observe how man is going through a scientific boom, an industrial boom, a population boom. A boom, period. Any day now he's going to have practical space ships. Meanwhile, he's also got the H-Bomb and the way he beats the drums on both sides of the Curtain, he's not against using it, if he could get away with it." Paul said, "I got it. So they're scared and are keeping an eye on us. That's an old one. I've read that a dozen times, dished up different." I shifted my shoulders. "Well, it's one possibility." "I got a better one. How's this. There's this alien life form that's way ahead of us. Their civilization is so old that they don't have any records of when it began and how it was in the early days. They've gone beyond things like wars and depressions and revolutions, and greed for power or any of these things giving us a bad time here on Earth. They're all like scholars, get it? And some of them are pretty jolly well taken by Earth, especially the way we are right now, with all the problems, get it? Things developing so fast we don't know where we're going or how we're going to get there." I finished my beer and clapped my hands for Mouley. "How do you mean, where we're going ?" "Well, take half the countries in the world today. They're trying to industrialize, modernize, catch up with the advanced countries. Look at Egypt, and Israel, and India and China, and Yugoslavia and Brazil, and all the rest. Trying to drag themselves up to the level of the advanced countries, and all using different methods of doing it. But look at the so-called advanced countries. Up to their bottoms in problems. Juvenile delinquents, climbing crime and suicide rates, the loony-bins full of the balmy, unemployed, threat of war, spending all their money on armaments instead of things like schools. All the bloody mess of it. Why, a man from Mars would be fascinated, like." Mouley came shuffling up in his babouche slippers and we both ordered another schooner of beer. Paul said seriously, "You know, there's only one big snag in this sort of talk. I've sorted the whole thing out before, and you always come up against this brick wall. Where are they, these observers, or scholars, or spies or whatever they are? Sooner or later we'd nab one of them. You know, Scotland Yard, or the F.B.I., or Russia's secret police, or the French Sûreté, or Interpol. This world is so deep in police, counter-espionage outfits and security agents that an alien would slip up in time, no matter how much he'd been trained. Sooner or later, he'd slip up, and they'd nab him." I shook my head. "Not necessarily. The first time I ever considered this possibility, it seemed to me that such an alien would base himself in London or New York. Somewhere where he could use the libraries for research, get the daily newspapers and the magazines. Be right in the center of things. But now I don't think so. I think he'd be right here in Tangier." "Why Tangier?" "It's the one town in the world where anything goes. Nobody gives a damn about you or your affairs. For instance, I've known you a year or more now, and I haven't the slightest idea of how you make your living." "That's right," Paul admitted. "In this town you seldom even ask a man where's he's from. He can be British, a White Russian, a Basque or a Sikh and nobody could care less. Where are you from, Rupert?" "California," I told him. "No, you're not," he grinned. I was taken aback. "What do you mean?" "I felt your mind probe back a few minutes ago when I was talking about Scotland Yard or the F.B.I. possibly flushing an alien. Telepathy is a sense not trained by the humanoids. If they had it, your job—and mine—would be considerably more difficult. Let's face it, in spite of these human bodies we're disguised in, neither of us is humanoid. Where are you really from, Rupert?" "Aldebaran," I said. "How about you?" "Deneb," he told me, shaking. We had a laugh and ordered another beer. "What're you doing here on Earth?" I asked him. "Researching for one of our meat trusts. We're protein eaters. Humanoid flesh is considered quite a delicacy. How about you?" "Scouting the place for thrill tourists. My job is to go around to these backward cultures and help stir up inter-tribal, or international, conflicts—all according to how advanced they are. Then our tourists come in—well shielded, of course—and get their kicks watching it." Paul frowned. "That sort of practice could spoil an awful lot of good meat." THE END Transcriber's Note: This etext was produced from Amazing Stories December 1960. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed. Minor spelling and typographical errors have been corrected without note. | D. The Earth has so many intelligence agencies, at least one would be watching when an alien gave itself away. |
What datasets were used for this work? | ### Introduction
The extraction of temporal relations among events is an important natural language understanding (NLU) task that can benefit many downstream tasks such as question answering, information retrieval, and narrative generation. The task can be modeled as building a graph for a given text, whose nodes represent events and edges are labeled with temporal relations correspondingly. Figure FIGREF1 illustrates such a graph for the text shown therein. The nodes assassination, slaughtered, rampage, war, and Hutu are the candidate events, and different types of edges specify different temporal relations between them: assassination is BEFORE rampage, rampage INCLUDES slaughtered, and the relation between slaughtered and war is VAGUE. Since “Hutu” is actually not an event, a system is expected to annotate the relations between “Hutu” and all other nodes in the graph as NONE (i.e., no relation). As far as we know, all existing systems treat this task as a pipeline of two separate subtasks, i.e., event extraction and temporal relation classification, and they also assume that gold events are given when training the relation classifier BIBREF0, BIBREF1, BIBREF2, BIBREF3, BIBREF4, BIBREF5. Specifically, they built end-to-end systems that extract events first and then predict temporal relations between them (Fig. FIGREF1). In these pipeline models, event extraction errors will propagate to the relation classification step and cannot be corrected afterwards. Our first contribution is the proposal of a joint model that extracts both events and temporal relations simultaneously (see Fig. FIGREF1). The motivation is that if we train the relation classifier with NONE relations between non-events, then it will potentially have the capability of correcting event extraction mistakes. For instance in Fig. FIGREF1, if the relation classifier predicts NONE for (Hutu, war) with a high confidence, then this is a strong signal that can be used by the event classifier to infer that at least one of them is not an event. Our second contribution is that we improve event representations by sharing the same contextualized embeddings and neural representation learner between the event extraction and temporal relation extraction modules for the first time. On top of the shared embeddings and neural representation learner, the proposed model produces a graph-structured output representing all the events and relations in the given sentences. A valid graph prediction in this context should satisfy two structural constraints. First, the temporal relation should always be NONE between two non-events or between one event and one non-event. Second, for those temporal relations among events, no loops should exist due to the transitive property of time (e.g., if A is before B and B is before C, then A must be before C). The validity of a graph is guaranteed by solving an integer linear programming (ILP) optimization problem with those structural constraints, and our joint model is trained by structural support vector machines (SSVM) in an end-to-end fashion. Results show that, according to the end-to-end $F_1$ score for temporal relation extraction, the proposed method improves CAEVO BIBREF3 by 10% on TB-Dense, and improves CogCompTime BIBREF6 by 6.8% on MATRES. We further show ablation studies to confirm that the proposed joint model with shared representations and structured learning is very effective for this task. ### Related Work
In this section we briefly summarize the existing work on event extraction and temporal relation extraction. To the best of our knowledge, there is no prior work on joint event and relation extraction, so we will review joint entity and relation extraction works instead. Existing event extraction methods in the temporal relation domain, as in the TempEval3 workshop BIBREF2, all use conventional machine learning models (logistic regression, SVM, or Max-entropy) with hand-engineered features (e.g., ClearTK BIBREF7 and NavyTime BIBREF8). While other domains have shown progress on event extraction using neural methods BIBREF9, BIBREF10, BIBREF11, recent progress in the temporal relation domain is focused more on the setting where gold events are provided. Therefore, we first show the performance of a neural event extractor on this task, although it is not our main contribution. Early attempts on temporal relation extraction use local pair-wise classification with hand-engineered features BIBREF12, BIBREF0, BIBREF13, BIBREF14. Later efforts, such as ClearTK BIBREF7, UTTime BIBREF15, NavyTime BIBREF8, and CAEVO BIBREF3 improve earlier work with better linguistic and syntactic rules. BIBREF16, BIBREF4, BIBREF17 explore structured learning for this task, and more recently, neural methods have also been shown effective BIBREF18, BIBREF19, BIBREF20, BIBREF5. In practice, we need to extract both events and those temporal relations among them from raw text. All the works above treat this as two subtasks that are solved in a pipeline. To the best of our knowledge, there has been no existing work on joint event-temporal relation extraction. However, the idea of “joint” has been studied for entity-relation extraction in many works. BIBREF21 frame their joint model as table filling tasks, map tabular representation into sequential predictions with heuristic rules, and construct global loss to compute the best joint predictions. BIBREF22 define a global structure for joint entity and relation extraction, encode local and global features based on domain and linguistic knowledge. and leverage beam-search to find global optimal assignments for entities and relations. BIBREF23 leverage LSTM architectures to jointly predict both entity and relations, but fall short on ensuring prediction consistency. BIBREF24 combine the benefits of both neural net and global optimization with beam search. Motivated by these works, we propose an end-to-end trainable neural structured support vector machine (neural SSVM) model to simultaneously extract events and their relations from text and ensure the global structure via ILP constraints. Next, we will describe in detail our proposed method. ### Joint Event-Relation Extraction Model
In this section we first provide an overview of our neural SSVM model, and then describe each component in our framework in detail (i.e., the multi-tasking neural scoring module, and how inference and learning are performed). We denote the set of all possible relation labels (including NONE) as $\mathcal {R}$, all event candidates (both events and non-events) as $\mathcal {E}$, and all relation candidates as $\mathcal {E}\mathcal {E}$. ### Joint Event-Relation Extraction Model ::: Neural SSVM
Our neural SSVM adapts the SSVM loss as: where $\bar{S}^n_{\mathcal {E}} = S(\hat{y}^n_\mathcal {E}; x^n) - S(y^n_\mathcal {E};x^n)$ and $\bar{S}^n_{\mathcal {R}} = S(\hat{y}^n_\mathcal {R}; x^n) - S(y^n_\mathcal {R};x^n)$ ; $\Phi $ denotes model parameters, $n$ indexes instances, $M^n = |\mathcal {E}|^n + |\mathcal {E}\mathcal {E}|^n$ denotes the total number of relations $|\mathcal {E}|^n$ and events $|\mathcal {E}\mathcal {E}|^n$ in instance $n$. $y^n,\hat{y}^n$ denote the gold and predicted global assignments of events and relations for instance $n$—each of which consists of either one hot vector representing true and predicted relation labels $y_{\mathcal {R}}^n, \hat{y}_{\mathcal {R}}^n \in \lbrace 0, 1\rbrace ^{|\mathcal {E}\mathcal {E}|}$, or entity labels $y_{\mathcal {E}}^n, \hat{y}_{\mathcal {E}}^n \in \lbrace 0, 1\rbrace ^{|\mathcal {E}|}$. A maximum a posteriori probability (MAP) inference is needed to find $\hat{y}^n$, which we formulate as an interger linear programming (ILP) problem and describe more details in Section SECREF12. $\Delta (y^n, \hat{y}^n)$ is a distance measurement between the gold and the predicted assignments; we simply use the Hamming distance. $C$ and $C_{\mathcal {E}}$ are the hyper-parameters to balance the losses between event, relation and the regularizer, and $S(y^n_\mathcal {E};x^n), S(y^n_\mathcal {R};x^n)$ are scoring functions, which we design a multi-tasking neural architecture to learn. The intuition behind the SSVM loss is that it requires the score of gold output structure $y^n$ to be greater than the score of the best output structure under the current model $\hat{y}^n$ with a margin $\Delta (y^n, \hat{y}^n)$ or else there will be some loss. The training objective is to minimize the loss. The major difference between our neural-SSVM and the traditional SSVM model is the scoring function. Traditional SSVM uses a linear function over hand-crafted features to compute the scores, whereas we propose to use a recurrent neural network to estimate the scoring function and train the entire architecture end-to-end. ### Joint Event-Relation Extraction Model ::: Multi-Tasking Neural Scoring Function
The recurrent neural network (RNN) architecture has been widely adopted by prior temporal extraction work to encode context information BIBREF18, BIBREF19, BIBREF20. Motivated by these works, we adopt a RNN-based scoring function for both event and relation prediction in order to learn features in a data driven way and capture long-term contexts in the input. In Fig. FIGREF6, we skip the input layer for simplicity. The bottom layer corresponds to contextualized word representations denoted as $v_k$. We use ($i, j$) $\in \mathcal {E}\mathcal {E}$ to denote a candidate relation and $i \in \mathcal {E}$ to indicate a candidate event in the input sentences of length N. We fix word embeddings computed by a pre-trained BERT-base model BIBREF27. They are then fed into a BiLSTM layer to further encode task-specific contextual information. Both event and relation tasks share this layer. The event scorer is illustrated by the left two branches following the BiLSTM layer. We simply concatenate both forward and backward hidden vectors to encode the context of each token. As for the relation scorer shown in the right branches, for each pair ($i,j$) we take the forward and backward hidden vectors corresponding to them, $f_i, b_i, f_j, b_j$, and concatenate them with linguistic features as in previous event relation prediction research. We denote linguistic features as $L_{i,j}$ and only use simple features provided in the original datasets: token distance, tense, and polarity of events. Finally, all hidden vectors and linguistic features are concatenated to form the input to compute the probability of being an event or a softmax distribution over all possible relation labels—which we refer to as the RNN-based scoring function in the following sections. ### Joint Event-Relation Extraction Model ::: MAP Inference
A MAP inference is needed both during training to obtain $\hat{y}^n$ in the loss function (Equation DISPLAY_FORM8), as well as during the test time to get globally coherent assignments. We formulate the inference problem as an ILP problem. The inference framework is established by constructing a global objective function using scores from local scorers and imposing several global constraints: 1) one-label assignment, 2) event-relation consistency, and 3) symmetry and transitivity as in BIBREF28, BIBREF29, BIBREF30, BIBREF31, BIBREF4. ### Joint Event-Relation Extraction Model ::: MAP Inference ::: Objective Function
The objective function of the global inference is to find the global assignment that has the highest probability under the current model, as specified in Equation DISPLAY_FORM14: where $y^e_k$ is a binary indicator of whether the $k$-th candidate is an event or not, and $y^r_{i,j}$ is a binary indicator specifying whether the global prediction of the relation between $(i,j)$ is $r \in \mathcal {R}$. $S(y^e_k,x), \forall e \in \lbrace 0, 1\rbrace $ and $S(y^r_{i,j},x), \forall r \in \mathcal {R}$ are the scoring functions obtained from the event and relation scoring functions, respectively. The output of the global inference $\bf {\hat{y}}$ is a collection of optimal label assignments for all events and relation candidates in a fixed context. $C_{\mathcal {E}}$ is a hyper-parameter controlling weights between relation and event. The constraint that follows immediately from the objective function is that the global inference should only assign one label for all entities and relations. ### Joint Event-Relation Extraction Model ::: MAP Inference ::: Constraints
We introduce several additional constraints to ensure the resulting optimal output graph forms a valid and plausible event graph. ### Joint Event-Relation Extraction Model ::: MAP Inference ::: Constraints ::: Event-Relation Consistency.
Event and relation prediction consistency is defined with the following property: a pair of input tokens have a positive temporal relation if and only if both tokens are events. The following global constraints will satisfy this property, where $e^P_i$ denotes an event and $e^N_i$ denotes a non-event token. $r^P_{i,j}$ indicates positive relations: BEFORE, AFTER, SIMULTANEOUS, INCLUDES, IS_INCLUDED, VAGUE and $r^N_{i,j}$ indicate a negative relation, i.e., NONE. A formal proof of this property can be found in Appendix A. ### Joint Event-Relation Extraction Model ::: MAP Inference ::: Constraints ::: Symmetry and Transitivity Constraint.
We also explore the symmetry and transitivity constraints of relations. They are specified as follows: Intuitively, the symmetry constraint forces two pairs of events with flipping orders to have reversed relations. For example, if $r_{i,j}$ = BEFORE, then $r_{j,i}$ = AFTER. The transitivity constraint rules that if ($i,j$), ($j,k$) and ($i,k$) pairs exist in the graph, the label (relation) prediction of ($i,k$) pair has to fall into the transitivity set specifyed by ($i,j$) and ($j,k$) pairs. The full transitivity table can be found in BIBREF25. ### Joint Event-Relation Extraction Model ::: Learning
We begin by experimenting with optimizing SSVM loss directly, but model performance degrades. Therefore, we develop a two-state learning approach which first trains a pipeline version of the joint model without feedback from global constraints. In other words, the local neural scoring functions are optimized with cross-entropy loss using gold events and relation candidates that are constructed directly from the outputs of the event model. During the second stage, we switch to the global SSVM loss function in Equation DISPLAY_FORM8 and re-optimize the network to adjust for global properties. We will provide more details in Section SECREF4. ### Implementation Details
In this section we describe implementation details of the baselines and our four models to build an end-to-end event temporal relation extraction system with an emphasis on the structured joint model. In Section SECREF6 we will compare and contrast them and show why our proposed structured joint model works the best. ### Implementation Details ::: Baselines
We run two event and relation extraction systems, CAEVO BIBREF3 and CogCompTime BIBREF6, on TB-Dense and MATRES, respectively. These two methods both leverage conventional learning algorithms (i.e., MaxEnt and averaged perceptron, respectively) based on manually designed features to obtain separate models for events and temporal relations, and conduct end-to-end relation extraction as a pipeline. Note BIBREF3 does not report event and end-to-end temporal relation extraction performances, so we calculate the scores per our implementation. ### Implementation Details ::: End-to-End Event Temporal Relation Extraction ::: Single-Task Model.
The most basic way to build an end-to-end system is to train separate event detection and relation prediction models with gold labels, as we mentioned in our introduction. In other words, the BiLSTM layer is not shared as in Fig. FIGREF6. During evaluation and test time, we use the outputs from the event detection model to construct relation candidates and apply the relation prediction model to make the final prediction. ### Implementation Details ::: End-to-End Event Temporal Relation Extraction ::: Multi-Task Model.
This is the same as the single-task model except that the BiLSTM layer is now shared for both event and relation tasks. Note that both single-task and multi-task models are not trained to tackle the NONE relation directly. They both rely on the predictions of the event model to annotate relations as either positive pairs or NONE. ### Implementation Details ::: End-to-End Event Temporal Relation Extraction ::: Pipeline Joint Model.
This shares the same architecture as the multi-task model, except that during training, we use the predictions of the event model to construct relation candidates to train the relation model. This strategy will generate NONE pairs during training if one argument of the relation candidate is not an event. These NONE pairs will help the relation model to distinguish negative relations from positive ones, and thus become more robust to event prediction errors. We train this model with gold events and relation candidates during the first several epochs in order to obtain a relatively accurate event model and switch to a pipeline version afterwards inspired by BIBREF23. ### Implementation Details ::: End-to-End Event Temporal Relation Extraction ::: Structured Joint Model.
This is described in detail in Section SECREF3. However, we experience difficulties in training the model with SSVM loss from scratch. This is due to large amounts of non-event tokens, and the model is not capable of distinguishing them in the beginning. We thus adopt a two-stage learning procedure where we take the best pipeline joint model and re-optimize it with the SSVM loss. To restrict the search space for events in the ILP inference of the SSVM loss, we use the predicted probabilities from the event detection model to filter out non-events since the event model has a strong performance, as shown in Section SECREF6. Note that this is very different from the pipeline model where events are first predicted and relations are constructed with predicted events. Here, we only leverage an additional hyper-parameter $T_{evt}$ to filter out highly unlikely event candidates. Both event and relation labels are assigned simutaneously during the global inference with ILP, as specified in Section SECREF12. We also filter out tokens with POS tags that do not appear in the training set as most of the events are either nouns or verbs in TB-Dense, and all events are verbs in MATRES. ### Implementation Details ::: End-to-End Event Temporal Relation Extraction ::: Hyper-Parameters.
All single-task, multi-task and pipeline joint models are trained by minimizing cross-entropy loss. We observe that model performances vary significantly with dropout ratio, hidden layer dimensions of the BiLSTM model and entity weight in the loss function (with relation weight fixed at 1.0). We leverage a pre-trained BERT model to compute word embedding and all MLP scoring functions have one hidden layer. In the SSVM loss function, we fix the value of $C = 1$, but fine-tune $C_\mathcal {E}$ in the objective function in Equation DISPLAY_FORM14. Hyper-parameters are chosen using a standard development set for TB-Dense and a random holdout-set based on an 80/20 split of training data for MATRES. To solve ILP in the inference process, we leverage an off-the-shelf solver provided by Gurobi optimizer; i.e. the best solutions from the Gurobi optimizer are inputs to the global training. The best combination of hyper-parameters can be found in Table 9 in our appendix. ### Experimental Setup
In this section we first provide a brief overview of temporal relation data and describe the specific datasets used in this paper. We also explain the evaluation metrics at the end. ### Experimental Setup ::: Temporal Relation Data
Temporal relation corpora such as TimeBank BIBREF32 and RED BIBREF33 facilitate the research in temporal relation extraction. The common issue in these corpora is missing annotations. Collecting densely annotated temporal relation corpora with all events and relations fully annotated is reported to be a challenging task as annotators could easily overlook some facts BIBREF34, BIBREF35, BIBREF3, BIBREF4, which made both modeling and evaluation extremely difficult in previous event temporal relation research. The TB-Dense dataset mitigates this issue by forcing annotators to examine all pairs of events within the same or neighboring sentences, and it has been widely evaluated on this task BIBREF3, BIBREF4, BIBREF19, BIBREF5. Recent data construction efforts such as MATRES BIBREF25 further enhance the data quality by using a multi-axis annotation scheme and adopting a start-point of events to improve inter-annotator agreements. We use TB-Dense and MATRES in our experiments and briefly summarize the data statistics in Table TABREF33. ### Experimental Setup ::: Evaluation Metrics
To be consistent with previous research, we adopt two different evaluation metrics. The first one is the standard micro-average scores. For densely annotated data, the micro-average metric should share the same precision, recall and F1 scores. However, since our joint model includes NONE pairs, we follow the convention of IE tasks and exclude them from evaluation. The second one is similar except that we exclude both NONE and VAGUE pairs following BIBREF6. Please refer to Figure 4 in the appendix for a visualizations of the two metrics. ### Results and Analysis
The main results of this paper can be found in Table TABREF34. All best-recall and F1 scores are achieved by our structured joint model, and the results outperform the baseline systems by 10.0% and 6.8% on end-to-end relation extraction per F1 scores and 3.5% and 2.6% on event extraction per F1 scores. The best precision score for the TB-Dense dataset is achieved by CAEVO, which indicates that the linguistic rule-based system can make highly precise predictions by being conservative. Table TABREF35 shows a more detailed analysis, in which we can see that our single-task models with BERT embeddings and a BiLSTM encoder already outperform the baseline systems on end-to-end relation extraction tasks by 4.9% and 4.4% respectively. In the following sections we discuss step-by-step improvement by adopting multi-task, pipeline joint, and structured joint models on end-to-end relation extraction, event extraction, and relation extraction on gold event pairs. ### Results and Analysis ::: End-to-End Relation Extraction ::: TB-Dense.
The improvements over the single-task model per F1 score are 4.1% and 4.2% for the multi-task and pipeline joint model respectively. This indicates that the pipeline joint model is helpful only marginally. Table TABREF46 shows that the structured joint model improves both precision and recall scores for BEFORE and AFTER and achieves the best end-to-end relation extraction performance at 49.4%—which outperforms the baseline system by 10.0% and the single-task model by 5.1%. ### Results and Analysis ::: End-to-End Relation Extraction ::: MATRES.
Compared to the single-task model, the multi-task model improves F1 scores by 1.5%, while the pipeline joint model improves F1 scores by 1.3%—which means that pipeline joint training does not bring any gains for MATRES. The structured joint model reaches the best end-to-end F1 score at 59.6%, which outperforms the baseline system by 6.8% and the single-task model by 2.4%. We speculate that the gains come from the joint model's ability to help deal with NONE pairs, since recall scores for BEFORE and AFTER increase by 1.5% and 1.1% respectively (Table 10 in our appendix). ### Results and Analysis ::: Event Extraction ::: TB-Dense.
Our structured joint model out-performs the CAEVO baseline by 3.5% and the single-task model by 1.3%. Improvements on event extraction can be difficult because our single-task model already works quite well with a close-to 89% F1 score, while the inter-annotator agreement for events in TimeBank documents is merely 87% BIBREF2. ### Results and Analysis ::: Event Extraction ::: MATRES.
The structured model outperforms the the baseline model and the single-task model by 2.6% and 0.9% respectively. However, we observe that the multi-task model has a slight drop in event extraction performance over the single-task model (86.4% vs. 86.9%). This indicates that incorporating relation signals are not particularly helpful for event extraction on MATRES. We speculate that one of the reasons could be the unique event characteristics in MATERS. As we described in Section SECREF32, all events in MATRES are verbs. It is possible that a more concentrated single-task model works better when events are homogeneous, whereas a multi-task model is more powerful when we have a mixture of event types, e.g., both verbs and nouns as in TB-Dense. ### Results and Analysis ::: Relation Extraction with Gold Events ::: TB-Dense.
There is much prior work on relation extraction based on gold events in TB-Dense. meng2018context proposed a neural model with global information that achieved the best results as far as we know. The improvement of our single-task model over that baseline is mostly attributable to the adoption of BERT embedding. We show that sharing the LSTM layer for both events and relations can help further improve performance of the relation classification task by 2.6%. For the joint models, since we do not train them on gold events, the evaluation would be meaningless. We simply skip this evaluation. ### Results and Analysis ::: Relation Extraction with Gold Events ::: MATRES.
Both single-task and multi-task models outperform the baseline by nearly 10%, while the improvement of multi-task over single task is marginal. In MATRES, a relation pair is equivalent to a verb pair, and thus the event prediction task probably does not provide much more information for relation extraction. In Table TABREF46 we further show the breakdown performances for each positive relation on TB-Dense. The breakdown on MATRES is shown in Table 10 in the appendix. BEFORE, AFTER and VAGUE are the three dominant label classes in TB-Dense. We observe that the linguistic rule-based model, CAEVO, tends to have a more evenly spread-out performance, whereas our neural network-based models are more likely to have concentrated predictions due to the imbalance of the training sample across different label classes. ### Results and Analysis ::: Discussion ::: Label Imbalance.
One way to mitigate the label imbalance issue is to increase the sample weights for small classes during model training. We investigate the impact of class weights by refitting our single-task model with larger weights on INCLUDES, IS_INCLUDED and VAGUE in the cross-entropy loss. Figure FIGREF50 shows that increasing class weights up to 4 times can significantly improve the F1 scores of INCLUDES and IS_INCLUDED classes with a decrease less than 2% for the overall F1 score. Performance of INCLUDES and IS_INCLUDED eventually degrades when class weights are too large. These results seem to suggest that more labels are needed in order to improve the performance on both of these two classes and the overall model. For SIMULTANEOUS, our model does not make any correct predictions for both TB-DENSE and MATRES by increasing class weight up to 10 times, which implies that SIMULTANEOUS could be a hard temporal relation to predict in general. ### Results and Analysis ::: Discussion ::: Global Constraints.
In Table TABREF51 we conduct an ablation study to understand the contributions from the event-relation prediction consistency constraint and the temporal relation transitivity constraint for the structured joint model. As we can see, the event-relation consistency help s improve the F1 scores by 0.9% and 1% for TB-Dense and MATRES, respectively, but the gain by using transitivity is either non-existing or marginal. We hypothesize two potential reasons: 1) We leveraged BERT contextualized embedding as word representation, which could tackle transitivity in the input context; 2) NONE pairs could make transitivity rule less useful, as positive pairs can be predicted as NONE and transitivity rule does not apply to NONE pairs. ### Results and Analysis ::: Discussion ::: Error Analysis.
By comparing gold and predicted labels for events and temporal relations and examining predicted probabilities for events, we identified three major sources of mistakes made by our structured model, as illustrated in Table TABREF57 with examples. ### Results and Analysis ::: Discussion ::: Type 1.
Both events in Ex 1 are assigned low scores by the event module ($<< 0.01$). Although the structured joint model is designed to predict events and relations jointly, we leverage the event module to filter out tokens with scores lower than a threshold. Consequently, some true events can be mistakenly predicted as non-events, and the relation pairs including them are automatically assigned NONE. ### Results and Analysis ::: Discussion ::: Type 2.
In Ex 2 the event module assigns high scores to tokens happened (0.97) and according (0.89), but according is not an event. When the structured model makes inference jointly, the decision will weigh heavily towards assigning 1 (event) to both tokens. With the event-relation consistency constraint, this pair is highly likely to be predicted as having a positive temporal relation. Nearly all mistakes made in this category follow the same pattern illustrated by this example. ### Results and Analysis ::: Discussion ::: Type 3.
The existence of VAGUE makes temporal relation prediction challenging as it can be easily confused with other temporal relations, as shown in Ex 3. This challenge is compounded with NONE in our end-to-end extraction task. Type 1 and Type 2 errors suggest that building a stronger event detection module will be helpful for both event and temporal relation extraction tasks. To improve the performance on VAGUE pairs, we could either build a stronger model that incorporates both contextual information and commonsense knowledge or create datasets with annotations that better separate VAGUE from other positive temporal relations. ### Conclusion
In this paper we investigate building an end-to-end event temporal relation extraction system. We propose a novel neural structured prediction model with joint representation learning to make predictions on events and relations simultaneously; this can avoid error propagation in previous pipeline systems. Experiments and comparative studies on two benchmark datasets show that the proposed model is effective for end-to-end event temporal relation extraction. Specifically, we improve the performances of previously published systems by 10% and 6.8% on the TB-Dense and MATRES datasets, respectively. Future research can focus on creating more robust structured constraints between events and relations, especially considering event types, to improve the quality of global assignments using ILP. Since a better event model is generally helpful for relation extraction, another promising direction would be to incorporate multiple datasets to enhance the performance of our event extraction systems. ### Acknowledgements
This work is supported in part by Contracts W911NF-15-1-0543 and HR0011-18-2-0052 with the US Defense Advanced Research Projects Agency (DARPA). Approved for Public Release, Distribution Unlimited. The views expressed are those of the authors and do not reflect the official policy or position of the Department of Defense or the U.S. Government. Figure 1: An illustration of event and relation models in our proposed joint framework. (a) is a (partial) graph of the output of the relation extraction model. “Hutu” is not an event and hence all relations including it should be annotated as NONE. (b) and (c) are comparisons between a pipeline model and our joint model. Figure 2: Deep neural network architecture for joint structured learning. Note that on the structured learning layer, grey bars denote tokens being predicted as events. Edge types between events follow the same notations as in 1a. yel = 0 (non-event), so all edges connecting to yel are NONE. y e i = 1, y e j = 1, y e k = 1 (events) and hence edges between them are forced to be the same (yrij = y r jk = y r ik = BEFORE in this example) by transitivity. These global assignments are input to compute the SSVM loss. Table 1: Data overview. Note that the numbers reported for MATRES do not include the AQUAINT section. Table 2: Event and Relation Extraction Results on TB-Dense and MATRES Table 3: Further ablation studies on event and relation extractions. Relation (G) denotes train and evaluate using gold events to compose relation candidates, whereas Relation (E) means end-to-end relation extraction. † is the event extraction and pipeline relation extraction F1 scores for CAEVO (Chambers et al., 2014). 57.0‡ is the best previously reported micro-average score for temporal relation extraction based on gold events by Meng and Rumshisky (2018). All MATRES baseline results are provided by Ning et al. (2018c). Table 4: Model performance breakdown for TB-Dense. “-” indicates no predictions were made for that particular label, probably due to the small size of the training sample. BEFORE (B), AFTER (A), INCLUDES (I), IS INCLUDED (II), SIMULTANEOUS (S), VAGUE (V) Table 5: Label Size Breakdown in the Test Data Figure 3: Performances from a single-task relation model under different class weights. Left-axis: overall model; right-axis: two minority relations. Table 6: Ablation Study on Global Constraints Table 7: Error Types Based on MATRES Test Data | TB-Dense, MATRES |
How is the accuracy of the system measured? | ### Introduction
In recent years, the spread of misinformation has become a growing concern for researchers and the public at large BIBREF1 . Researchers at MIT found that social media users are more likely to share false information than true information BIBREF2 . Due to renewed focus on finding ways to foster healthy political conversation, the profile of factcheckers has been raised. Factcheckers positively influence public debate by publishing good quality information and asking politicians and journalists to retract misleading or false statements. By calling out lies and the blurring of the truth, they make those in positions of power accountable. This is a result of labour intensive work that involves monitoring the news for spurious claims and carrying out rigorous research to judge credibility. So far, it has only been possible to scale their output upwards by hiring more personnel. This is problematic because newsrooms need significant resources to employ factcheckers. Publication budgets have been decreasing, resulting in a steady decline in the size of their workforce BIBREF0 . Factchecking is not a directly profitable activity, which negatively affects the allocation of resources towards it in for-profit organisations. It is often taken on by charities and philanthropists instead. To compensate for this shortfall, our strategy is to harness the latest developments in NLP to make factchecking more efficient and therefore less costly. To this end, the new field of automated factchecking has captured the imagination of both non-profits and start-ups BIBREF3 , BIBREF4 , BIBREF5 . It aims to speed up certain aspects of the factchecking process rather than create AI that can replace factchecking personnel. This includes monitoring claims that are made in the news, aiding decisions about which statements are the most important to check and automatically retrieving existing factchecks that are relevant to a new claim. The claim detection and claim clustering methods that we set out in this paper can be applied to each of these. We sought to devise a system that would automatically detect claims in articles and compare them to previously submitted claims. Storing the results to allow a factchecker's work on one of these claims to be easily transferred to others in the same cluster. ### Related Work
It is important to decide what sentences are claims before attempting to cluster them. The first such claim detection system to have been created is ClaimBuster BIBREF6 , which scores sentences with an SVM to determine how likely they are to be politically pertinent statements. Similarly, ClaimRank BIBREF7 uses real claims checked by factchecking institutions as training data in order to surface sentences that are worthy of factchecking. These methods deal with the question of what is a politically interesting claim. In order to classify the objective qualities of what set apart different types of claims, the ClaimBuster team created PolitiTax BIBREF8 , a taxonomy of claims, and factchecking organisation Full Fact BIBREF9 developed their preferred annotation schema for statements in consultation with their own factcheckers. This research provides a more solid framework within which to construct claim detection classifiers. The above considers whether or not a sentence is a claim, but often claims are subsections of sentences and multiple claims might be found in one sentence. In order to accommodate this, BIBREF10 proposes extracting phrases called Context Dependent Claims (CDC) that are relevant to a certain `Topic'. Along these lines, BIBREF11 proposes new definitions for frames to be incorporated into FrameNet BIBREF12 that are specific to facts, in particular those found in a political context. Traditional text clustering methods, using TFIDF and some clustering algorithm, are poorly suited to the problem of clustering and comparing short texts, as they can be semantically very similar but use different words. This is a manifestation of the the data sparsity problem with Bag-of-Words (BoW) models. BIBREF16 . Dimensionality reduction methods such as Latent Dirichlet Allocation (LDA) can help solve this problem by giving a dense approximation of this sparse representation BIBREF17 . More recently, efforts in this area have used text embedding-based systems in order to capture dense representation of the texts BIBREF18 . Much of this recent work has relied on the increase of focus in word and text embeddings. Text embeddings have been an increasingly popular tool in NLP since the introduction of Word2Vec BIBREF19 , and since then the number of different embeddings has exploded. While many focus on giving a vector representation of a word, an increasing number now exist that will give a vector representation of a entire sentence or text. Following on from this work, we seek to devise a system that can run online, performing text clustering on the embeddings of texts one at a time Some considerations to bear in mind when deciding on an embedding scheme to use are: the size of the final vector, the complexity of the model itself and, if using a pretrained implementation, the data the model has been trained on and whether it is trained in a supervised or unsupervised manner. The size of the embedding can have numerous results downstream. In our example we will be doing distance calculations on the resultant vectors and therefore any increase in length will increase the complexity of those distance calculations. We would therefore like as short a vector as possible, but we still wish to capture all salient information about the claim; longer vectors have more capacity to store information, both salient and non-salient. A similar effect is seen for the complexity of the model. A more complicated model, with more trainable parameters, may be able to capture finer details about the text, but it will require a larger corpus to achieve this, and will require more computational time to calculate the embeddings. We should therefore attempt to find the simplest embedding system that can accurately solve our problem. When attempting to use pretrained models to help in other areas, it is always important to ensure that the models you are using are trained on similar material, to increase the chance that their findings will generalise to the new problem. Many unsupervised text embeddings are trained on the CommonCrawl dataset of approx. 840 billion tokens. This gives a huge amount of data across many domains, but requires a similarly huge amount of computing power to train on the entire dataset. Supervised datasets are unlikely ever to approach such scale as they require human annotations which can be expensive to assemble. The SNLI entailment dataset is an example of a large open source dataset BIBREF20 . It features pairs of sentences along with labels specifying whether or not one entails the other. Google's Universal Sentence Encoder (USE) BIBREF14 is a sentence embedding created with a hybrid supervised/unsupervised method, leveraging both the vast amounts of unsupervised training data and the extra detail that can be derived from a supervised method. The SNLI dataset and the related MultiNLI dataset are often used for this because textual entailment is seen as a good basis for general Natural Language Understanding (NLU) BIBREF21 . ### Method
It is much easier to build a dataset and reliably evaluate a model if the starting definitions are clear and objective. Questions around what is an interesting or pertinent claim are inherently subjective. For example, it is obvious that a politician will judge their opponents' claims to be more important to factcheck than their own. Therefore, we built on the methodologies that dealt with the objective qualities of claims, which were the PolitiTax and Full Fact taxonomies. We annotated sentences from our own database of news articles based on a combination of these. We also used the Full Fact definition of a claim as a statement about the world that can be checked. Some examples of claims according to this definition are shown in Table TABREF3 . We decided the first statement was a claim since it declares the occurrence of an event, while the second was considered not to be a claim as it is an expression of feeling. Full Fact's approach centred around using sentence embeddings as a feature engineering step, followed by a simple classifier such as logistic regression, which is what we used. They used Facebook's sentence embeddings, InferSent BIBREF13 , which was a recent breakthrough at the time. Such is the speed of new development in the field that since then, several papers describing textual embeddings have been published. Due to the fact that we had already evaluated embeddings for clustering, and therefore knew our system would rely on Google USE Large BIBREF14 , we decided to use this instead. We compared this to TFIDF and Full Fact's results as baselines. The results are displayed in Table TABREF4 . However, ClaimBuster and Full Fact focused on live factchecking of TV debates. Logically is a news aggregator and we analyse the bodies of published news stories. We found that in our corpus, the majority of sentences are claims and therefore our model needed to be as selective as possible. In practice, we choose to filter out sentences that are predictions since generally the substance of the claim cannot be fully checked until after the event has occurred. Likewise, we try to remove claims based on personal experience or anecdotal evidence as they are difficult to verify. ### Choosing an embedding
In order to choose an embedding, we sought a dataset to represent our problem. Although no perfect matches exist, we decided upon the Quora duplicate question dataset BIBREF22 as the best match. To study the embeddings, we computed the euclidean distance between the two questions using various embeddings, to study the distance between semantically similar and dissimilar questions. The graphs in figure 1 show the distances between duplicate and non-duplicate questions using different embedding systems. The X axis shows the euclidean distance between vectors and the Y axis frequency. A perfect result would be a blue peak to the left and an entirely disconnected orange spike to the right, showing that all non-duplicate questions have a greater euclidean distance than the least similar duplicate pair of questions. As can be clearly seen in the figure above, Elmo BIBREF23 and Infersent BIBREF13 show almost no separation and therefore cannot be considered good models for this problem. A much greater disparity is shown by the Google USE models BIBREF14 , and even more for the Google USE Large model. In fact the Google USE Large achieved a F1 score of 0.71 for this task without any specific training, simply by choosing a threshold below which all sentence pairs are considered duplicates. In order to test whether these results generalised to our domain, we devised a test that would make use of what little data we had to evaluate. We had no original data on whether sentences were semantically similar, but we did have a corpus of articles clustered into stories. Working on the assumption that similar claims would be more likely to be in the same story, we developed an equation to judge how well our corpus of sentences was clustered, rewarding clustering which matches the article clustering and the total number of claims clustered. The precise formula is given below, where INLINEFORM0 is the proportion of claims in clusters from one story cluster, INLINEFORM1 is the proportion of claims in the correct claim cluster, where they are from the most common story cluster, and INLINEFORM2 is the number of claims placed in clusters. A,B and C are parameters to tune. INLINEFORM3 figureFormula to assess the correctness of claim clusters based on article clusters This method is limited in how well it can represent the problem, but it can give indications as to a good or bad clustering method or embedding, and can act as a check that the findings we obtained from the Quora dataset will generalise to our domain. We ran code which vectorized 2,000 sentences and then used the DBScan clustering method BIBREF24 to cluster using a grid search to find the best INLINEFORM0 value, maximizing this formula. We used DBScan as it mirrored the clustering method used to derive the original article clusters. The results for this experiment can be found in Table TABREF10 . We included TFIDF in the experiment as a baseline to judge other results. It is not suitable for our eventual purposes, but it the basis of the original keyword-based model used to build the clusters . That being said, TFIDF performs very well, with only Google USE Large and Infersent coming close in terms of `accuracy'. In the case of Infersent, this comes with the penalty of a much smaller number of claims included in the clusters. Google USE Large, however, clusters a greater number and for this reason we chose to use Google's USE Large. Since Google USE Large was the best-performing embedding in both the tests we devised, this was our chosen embedding to use for clustering. However as can be seen from the results shown above, this is not a perfect solution and the inaccuracy here will introduce inaccuracy further down the clustering pipeline. ### Clustering Method
We decided to follow a methodology upon the DBScan method of clustering BIBREF24 . DBScan considers all distances between pairs of points. If they are under INLINEFORM0 then those two are linked. Once the number of connected points exceeds a minimum size threshold, they are considered a cluster and all other points are considered to be unclustered. This method is advantageous for our purposes because unlike other methods, such as K-Means, it does not require the number of clusters to be specified. To create a system that can build clusters dynamically, adding one point at a time, we set the minimum cluster size to one, meaning that every point is a member of a cluster. A potential disadvantage of this method is that because points require only one connection to a cluster to join it, they may only be related to one point in the cluster, but be considered in the same cluster as all of them. In small examples this is not a problem as all points in the cluster should be very similar. However as the number of points being considered grows, this behaviour raises the prospect of one or several borderline clustering decisions leading to massive clusters made from tenuous connections between genuine clusters. To mitigate this problem we used a method described in the Newslens paper BIBREF25 to solve a similar problem when clustering entire articles. We stored all of our claims in a graph with the connections between them added when the distance between them was determined to be less than INLINEFORM0 . To determine the final clusters we run a Louvain Community Detection BIBREF26 over this graph to split it into defined communities. This improved the compactness of a cluster. When clustering claims one by one, this algorithm can be performed on the connected subgraph featuring the new claim, to reduce the computation required. As this method involves distance calculations between the claim being added and every existing claim, the time taken to add one claim will increase roughly linearly with respect to the number of previous claims. Through much optimization we have brought the computational time down to approximately 300ms per claim, which stays fairly static with respect to the number of previous claims. ### Next Steps
The clustering described above is heavily dependent on the embedding used. The rate of advances in this field has been rapid in recent years, but an embedding will always be an imperfect representation of an claim and therefore always an area of improvement. A domain specific-embedding will likely offer a more accurate representation but creates problems with clustering claims from different domains. They also require a huge amount of data to give a good model and that is not possible in all domains. Table 1: Examples of claims taken from real articles. Table 2: Claim Detection Results. Figure 1: Analysis of Different Embeddings on the Quora Question Answering Dataset Table 3: Comparing Sentence Embeddings for Clustering News Claims. | F1 score of 0.71 for this task without any specific training, simply by choosing a threshold below which all sentence pairs are considered duplicates, distances between duplicate and non-duplicate questions using different embedding systems |
How is face and audio data analysis evaluated? | ### Introduction
Automatic emotion recognition is commonly understood as the task of assigning an emotion to a predefined instance, for example an utterance (as audio signal), an image (for instance with a depicted face), or a textual unit (e.g., a transcribed utterance, a sentence, or a Tweet). The set of emotions is often following the original definition by Ekman Ekman1992, which includes anger, fear, disgust, sadness, joy, and surprise, or the extension by Plutchik Plutchik1980 who adds trust and anticipation. Most work in emotion detection is limited to one modality. Exceptions include Busso2004 and Sebe2005, who investigate multimodal approaches combining speech with facial information. Emotion recognition in speech can utilize semantic features as well BIBREF0. Note that the term “multimodal” is also used beyond the combination of vision, audio, and text. For example, Soleymani2012 use it to refer to the combination of electroencephalogram, pupillary response and gaze distance. In this paper, we deal with the specific situation of car environments as a testbed for multimodal emotion recognition. This is an interesting environment since it is, to some degree, a controlled environment: Dialogue partners are limited in movement, the degrees of freedom for occurring events are limited, and several sensors which are useful for emotion recognition are already integrated in this setting. More specifically, we focus on emotion recognition from speech events in a dialogue with a human partner and with an intelligent agent. Also from the application point of view, the domain is a relevant choice: Past research has shown that emotional intelligence is beneficial for human computer interaction. Properly processing emotions in interactions increases the engagement of users and can improve performance when a specific task is to be fulfilled BIBREF1, BIBREF2, BIBREF3, BIBREF4. This is mostly based on the aspect that machines communicating with humans appear to be more trustworthy when they show empathy and are perceived as being natural BIBREF3, BIBREF5, BIBREF4. Virtual agents play an increasingly important role in the automotive context and the speech modality is increasingly being used in cars due to its potential to limit distraction. It has been shown that adapting the in-car speech interaction system according to the drivers' emotional state can help to enhance security, performance as well as the overall driving experience BIBREF6, BIBREF7. With this paper, we investigate how each of the three considered modalitites, namely facial expressions, utterances of a driver as an audio signal, and transcribed text contributes to the task of emotion recognition in in-car speech interactions. We focus on the five emotions of joy, insecurity, annoyance, relaxation, and boredom since terms corresponding to so-called fundamental emotions like fear have been shown to be associated to too strong emotional states than being appropriate for the in-car context BIBREF8. Our first contribution is the description of the experimental setup for our data collection. Aiming to provoke specific emotions with situations which can occur in real-world driving scenarios and to induce speech interactions, the study was conducted in a driving simulator. Based on the collected data, we provide baseline predictions with off-the-shelf tools for face and speech emotion recognition and compare them to a neural network-based approach for emotion recognition from text. Our second contribution is the introduction of transfer learning to adapt models trained on established out-of-domain corpora to our use case. We work on German language, therefore the transfer consists of a domain and a language transfer. ### Related Work ::: Facial Expressions
A common approach to encode emotions for facial expressions is the facial action coding system FACS BIBREF9, BIBREF10, BIBREF11. As the reliability and reproducability of findings with this method have been critically discussed BIBREF12, the trend has increasingly shifted to perform the recognition directly on images and videos, especially with deep learning. For instance, jung2015joint developed a model which considers temporal geometry features and temporal appearance features from image sequences. kim2016hierarchical propose an ensemble of convolutional neural networks which outperforms isolated networks. In the automotive domain, FACS is still popular. Ma2017 use support vector machines to distinguish happy, bothered, confused, and concentrated based on data from a natural driving environment. They found that bothered and confused are difficult to distinguish, while happy and concentrated are well identified. Aiming to reduce computational cost, Tews2011 apply a simple feature extraction using four dots in the face defining three facial areas. They analyze the variance of the three facial areas for the recognition of happy, anger and neutral. Ihme2018 aim at detecting frustration in a simulator environment. They induce the emotion with specific scenarios and a demanding secondary task and are able to associate specific face movements according to FACS. Paschero2012 use OpenCV (https://opencv.org/) to detect the eyes and the mouth region and track facial movements. They simulate different lightning conditions and apply a multilayer perceptron for the classification task of Ekman's set of fundamental emotions. Overall, we found that studies using facial features usually focus on continuous driver monitoring, often in driver-only scenarios. In contrast, our work investigates the potential of emotion recognition during speech interactions. ### Related Work ::: Acoustic
Past research on emotion recognition from acoustics mainly concentrates on either feature selection or the development of appropriate classifiers. rao2013emotion as well as ververidis2004automatic compare local and global features in support vector machines. Next to such discriminative approaches, hidden Markov models are well-studied, however, there is no agreement on which feature-based classifier is most suitable BIBREF13. Similar to the facial expression modality, recent efforts on applying deep learning have been increased for acoustic speech processing. For instance, lee2015high use a recurrent neural network and palaz2015analysis apply a convolutional neural network to the raw speech signal. Neumann2017 as well as Trigeorgis2016 analyze the importance of features in the context of deep learning-based emotion recognition. In the automotive sector, Boril2011 approach the detection of negative emotional states within interactions between driver and co-driver as well as in calls of the driver towards the automated spoken dialogue system. Using real-world driving data, they find that the combination of acoustic features and their respective Gaussian mixture model scores performs best. Schuller2006 collects 2,000 dialog turns directed towards an automotive user interface and investigate the classification of anger, confusion, and neutral. They show that automatic feature generation and feature selection boost the performance of an SVM-based classifier. Further, they analyze the performance under systematically added noise and develop methods to mitigate negative effects. For more details, we refer the reader to the survey by Schuller2018. In this work, we explore the straight-forward application of domain independent software to an in-car scenario without domain-specific adaptations. ### Related Work ::: Text
Previous work on emotion analysis in natural language processing focuses either on resource creation or on emotion classification for a specific task and domain. On the side of resource creation, the early and influential work of Pennebaker2015 is a dictionary of words being associated with different psychologically relevant categories, including a subset of emotions. Another popular resource is the NRC dictionary by Mohammad2012b. It contains more than 10000 words for a set of discrete emotion classes. Other resources include WordNet Affect BIBREF14 which distinguishes particular word classes. Further, annotated corpora have been created for a set of different domains, for instance fairy tales BIBREF15, Blogs BIBREF16, Twitter BIBREF17, BIBREF18, BIBREF19, BIBREF20, BIBREF21, Facebook BIBREF22, news headlines BIBREF23, dialogues BIBREF24, literature BIBREF25, or self reports on emotion events BIBREF26 (see BIBREF27 for an overview). To automatically assign emotions to textual units, the application of dictionaries has been a popular approach and still is, particularly in domains without annotated corpora. Another approach to overcome the lack of huge amounts of annotated training data in a particular domain or for a specific topic is to exploit distant supervision: use the signal of occurrences of emoticons or specific hashtags or words to automatically label the data. This is sometimes referred to as self-labeling BIBREF21, BIBREF28, BIBREF29, BIBREF30. A variety of classification approaches have been tested, including SNoW BIBREF15, support vector machines BIBREF16, maximum entropy classification, long short-term memory network, and convolutional neural network models BIBREF18. More recently, the state of the art is the use of transfer learning from noisy annotations to more specific predictions BIBREF29. Still, it has been shown that transferring from one domain to another is challenging, as the way emotions are expressed varies between areas BIBREF27. The approach by Felbo2017 is different to our work as they use a huge noisy data set for pretraining the model while we use small high quality data sets instead. Recently, the state of the art has also been pushed forward with a set of shared tasks, in which the participants with top results mostly exploit deep learning methods for prediction based on pretrained structures like embeddings or language models BIBREF21, BIBREF31, BIBREF20. Our work follows this approach and builds up on embeddings with deep learning. Furthermore, we approach the application and adaption of text-based classifiers to the automotive domain with transfer learning. ### Data set Collection
The first contribution of this paper is the construction of the AMMER data set which we describe in the following. We focus on the drivers' interactions with both a virtual agent as well as a co-driver. To collect the data in a safe and controlled environment and to be able to consider a variety of predefined driving situations, the study was conducted in a driving simulator. ### Data set Collection ::: Study Setup and Design
The study environment consists of a fixed-base driving simulator running Vires's VTD (Virtual Test Drive, v2.2.0) simulation software (https://vires.com/vtd-vires-virtual-test-drive/). The vehicle has an automatic transmission, a steering wheel and gas and brake pedals. We collect data from video, speech and biosignals (Empatica E4 to record heart rate, electrodermal activity, skin temperature, not further used in this paper) and questionnaires. Two RGB cameras are fixed in the vehicle to capture the drivers face, one at the sun shield above the drivers seat and one in the middle of the dashboard. A microphone is placed on the center console. One experimenter sits next to the driver, the other behind the simulator. The virtual agent accompanying the drive is realized as Wizard-of-Oz prototype which enables the experimenter to manually trigger prerecorded voice samples playing trough the in-car speakers and to bring new content to the center screen. Figure FIGREF4 shows the driving simulator. The experimental setting is comparable to an everyday driving task. Participants are told that the goal of the study is to evaluate and to improve an intelligent driving assistant. To increase the probability of emotions to arise, participants are instructed to reach the destination of the route as fast as possible while following traffic rules and speed limits. They are informed that the time needed for the task would be compared to other participants. The route comprises highways, rural roads, and city streets. A navigation system with voice commands and information on the screen keeps the participants on the predefined track. To trigger emotion changes in the participant, we use the following events: (i) a car on the right lane cutting off to the left lane when participants try to overtake followed by trucks blocking both lanes with a slow overtaking maneuver (ii) a skateboarder who appears unexpectedly on the street and (iii) participants are praised for reaching the destination unexpectedly quickly in comparison to previous participants. Based on these events, we trigger three interactions (Table TABREF6 provides examples) with the intelligent agent (Driver-Agent Interactions, D–A). Pretending to be aware of the current situation, e. g., to recognize unusual driving behavior such as strong braking, the agent asks the driver to explain his subjective perception of these events in detail. Additionally, we trigger two more interactions with the intelligent agent at the beginning and at the end of the drive, where participants are asked to describe their mood and thoughts regarding the (upcoming) drive. This results in five interactions between the driver and the virtual agent. Furthermore, the co-driver asks three different questions during sessions with light traffic and low cognitive demand (Driver-Co-Driver Interactions, D–Co). These questions are more general and non-traffic-related and aim at triggering the participants' memory and fantasy. Participants are asked to describe their last vacation, their dream house and their idea of the perfect job. In sum, there are eight interactions per participant (5 D–A, 3 D–Co). ### Data set Collection ::: Procedure
At the beginning of the study, participants were welcomed and the upcoming study procedure was explained. Subsequently, participants signed a consent form and completed a questionnaire to provide demographic information. After that, the co-driving experimenter started with the instruction in the simulator which was followed by a familiarization drive consisting of highway and city driving and covering different driving maneuvers such as tight corners, lane changing and strong braking. Subsequently, participants started with the main driving task. The drive had a duration of 20 minutes containing the eight previously mentioned speech interactions. After the completion of the drive, the actual goal of improving automatic emotional recognition was revealed and a standard emotional intelligence questionnaire, namely the TEIQue-SF BIBREF32, was handed to the participants. Finally, a retrospective interview was conducted, in which participants were played recordings of their in-car interactions and asked to give discrete (annoyance, insecurity, joy, relaxation, boredom, none, following BIBREF8) was well as dimensional (valence, arousal, dominance BIBREF33 on a 11-point scale) emotion ratings for the interactions and the according situations. We only use the discrete class annotations in this paper. ### Data set Collection ::: Data Analysis
Overall, 36 participants aged 18 to 64 years ($\mu $=28.89, $\sigma $=12.58) completed the experiment. This leads to 288 interactions, 180 between driver and the agent and 108 between driver and co-driver. The emotion self-ratings from the participants yielded 90 utterances labeled with joy, 26 with annoyance, 49 with insecurity, 9 with boredom, 111 with relaxation and 3 with no emotion. One example interaction per interaction type and emotion is shown in Table TABREF7. For further experiments, we only use joy, annoyance/anger, and insecurity/fear due to the small sample size for boredom and no emotion and under the assumption that relaxation brings little expressivity. ### Methods ::: Emotion Recognition from Facial Expressions
We preprocess the visual data by extracting the sequence of images for each interaction from the point where the agent's or the co-driver's question was completely uttered until the driver's response stops. The average length is 16.3 seconds, with the minimum at 2.2s and the maximum at 54.7s. We apply an off-the-shelf tool for emotion recognition (the manufacturer cannot be disclosed due to licensing restrictions). It delivers frame-by-frame scores ($\in [0;100]$) for discrete emotional states of joy, anger and fear. While joy corresponds directly to our annotation, we map anger to our label annoyance and fear to our label insecurity. The maximal average score across all frames constitutes the overall classification for the video sequence. Frames where the software is not able to detect the face are ignored. ### Methods ::: Emotion Recognition from Audio Signal
We extract the audio signal for the same sequence as described for facial expressions and apply an off-the-shelf tool for emotion recognition. The software delivers single classification scores for a set of 24 discrete emotions for the entire utterance. We consider the outputs for the states of joy, anger, and fear, mapping analogously to our classes as for facial expressions. Low-confidence predictions are interpreted as “no emotion”. We accept the emotion with the highest score as the discrete prediction otherwise. ### Methods ::: Emotion Recognition from Transcribed Utterances
For the emotion recognition from text, we manually transcribe all utterances of our AMMER study. To exploit existing and available data sets which are larger than the AMMER data set, we develop a transfer learning approach. We use a neural network with an embedding layer (frozen weights, pre-trained on Common Crawl and Wikipedia BIBREF36), a bidirectional LSTM BIBREF37, and two dense layers followed by a soft max output layer. This setup is inspired by BIBREF38. We use a dropout rate of 0.3 in all layers and optimize with Adam BIBREF39 with a learning rate of $10^{-5}$ (These parameters are the same for all further experiments). We build on top of the Keras library with the TensorFlow backend. We consider this setup our baseline model. We train models on a variety of corpora, namely the common format published by BIBREF27 of the FigureEight (formally known as Crowdflower) data set of social media, the ISEAR data BIBREF40 (self-reported emotional events), and, the Twitter Emotion Corpus (TEC, weakly annotated Tweets with #anger, #disgust, #fear, #happy, #sadness, and #surprise, Mohammad2012). From all corpora, we use instances with labels fear, anger, or joy. These corpora are English, however, we do predictions on German utterances. Therefore, each corpus is preprocessed to German with Google Translate. We remove URLs, user tags (“@Username”), punctuation and hash signs. The distributions of the data sets are shown in Table TABREF12. To adapt models trained on these data, we apply transfer learning as follows: The model is first trained until convergence on one out-of-domain corpus (only on classes fear, joy, anger for compatibility reasons). Then, the parameters of the bi-LSTM layer are frozen and the remaining layers are further trained on AMMER. This procedure is illustrated in Figure FIGREF13 ### Results ::: Facial Expressions and Audio
Table TABREF16 shows the confusion matrices for facial and audio emotion recognition on our complete AMMER data set and Table TABREF17 shows the results per class for each method, including facial and audio data and micro and macro averages. The classification from facial expressions yields a macro-averaged $\text{F}_1$ score of 33 % across the three emotions joy, insecurity, and annoyance (P=0.31, R=0.35). While the classification results for joy are promising (R=43 %, P=57 %), the distinction of insecurity and annoyance from the other classes appears to be more challenging. Regarding the audio signal, we observe a macro $\text{F}_1$ score of 29 % (P=42 %, R=22 %). There is a bias towards negative emotions, which results in a small number of detected joy predictions (R=4 %). Insecurity and annoyance are frequently confused. ### Results ::: Text from Transcribed Utterances
The experimental setting for the evaluation of emotion recognition from text is as follows: We evaluate the BiLSTM model in three different experiments: (1) in-domain, (2) out-of-domain and (3) transfer learning. For all experiments we train on the classes anger/annoyance, fear/insecurity and joy. Table TABREF19 shows all results for the comparison of these experimental settings. ### Results ::: Text from Transcribed Utterances ::: Experiment 1: In-Domain application
We first set a baseline by validating our models on established corpora. We train the baseline model on 60 % of each data set listed in Table TABREF12 and evaluate that model with 40 % of the data from the same domain (results shown in the column “In-Domain” in Table TABREF19). Excluding AMMER, we achieve an average micro $\text{F}_1$ of 68 %, with best results of F$_1$=73 % on TEC. The model trained on our AMMER corpus achieves an F1 score of 57%. This is most probably due to the small size of this data set and the class bias towards joy, which makes up more than half of the data set. These results are mostly in line with Bostan2018. ### Results ::: Text from Transcribed Utterances ::: Experiment 2: Simple Out-Of-Domain application
Now we analyze how well the models trained in Experiment 1 perform when applied to our data set. The results are shown in column “Simple” in Table TABREF19. We observe a clear drop in performance, with an average of F$_1$=48 %. The best performing model is again the one trained on TEC, en par with the one trained on the Figure8 data. The model trained on ISEAR performs second best in Experiment 1, it performs worst in Experiment 2. ### Results ::: Text from Transcribed Utterances ::: Experiment 3: Transfer Learning application
To adapt models trained on previously existing data sets to our particular application, the AMMER corpus, we apply transfer learning. Here, we perform leave-one-out cross validation. As pre-trained models we use each model from Experiment 1 and further optimize with the training subset of each crossvalidation iteration of AMMER. The results are shown in the column “Transfer L.” in Table TABREF19. The confusion matrix is also depicted in Table TABREF16. With this procedure we achieve an average performance of F$_1$=75 %, being better than the results from the in-domain Experiment 1. The best performance of F$_1$=76 % is achieved with the model pre-trained on each data set, except for ISEAR. All transfer learning models clearly outperform their simple out-of-domain counterpart. To ensure that this performance increase is not only due to the larger data set, we compare these results to training the model without transfer on a corpus consisting of each corpus together with AMMER (again, in leave-one-out crossvalidation). These results are depicted in column “Joint C.”. Thus, both settings, “transfer learning” and “joint corpus” have access to the same information. The results show an increase in performance in contrast to not using AMMER for training, however, the transfer approach based on partial retraining the model shows a clear improvement for all models (by 7pp for Figure8, 10pp for EmoInt, 8pp for TEC, 13pp for ISEAR) compared to the ”Joint” setup. ### Summary & Future Work
We described the creation of the multimodal AMMER data with emotional speech interactions between a driver and both a virtual agent and a co-driver. We analyzed the modalities of facial expressions, acoustics, and transcribed utterances regarding their potential for emotion recognition during in-car speech interactions. We applied off-the-shelf emotion recognition tools for facial expressions and acoustics. For transcribed text, we developed a neural network-based classifier with transfer learning exploiting existing annotated corpora. We find that analyzing transcribed utterances is most promising for classification of the three emotional states of joy, annoyance and insecurity. Our results for facial expressions indicate that there is potential for the classification of joy, however, the states of annoyance and insecurity are not well recognized. Future work needs to investigate more sophisticated approaches to map frame predictions to sequence predictions. Furthermore, movements of the mouth region during speech interactions might negatively influence the classification from facial expressions. Therefore, the question remains how facial expressions can best contribute to multimodal detection in speech interactions. Regarding the classification from the acoustic signal, the application of off-the-shelf classifiers without further adjustments seems to be challenging. We find a strong bias towards negative emotional states for our experimental setting. For instance, the personalization of the recognition algorithm (e. g., mean and standard deviation normalization) could help to adapt the classification for specific speakers and thus to reduce this bias. Further, the acoustic environment in the vehicle interior has special properties and the recognition software might need further adaptations. Our transfer learning-based text classifier shows considerably better results. This is a substantial result in its own, as only one previous method for transfer learning in emotion recognition has been proposed, in which a sentiment/emotion specific source for labels in pre-training has been used, to the best of our knowledge BIBREF29. Other applications of transfer learning from general language models include BIBREF41, BIBREF42. Our approach is substantially different, not being trained on a huge amount of noisy data, but on smaller out-of-domain sets of higher quality. This result suggests that emotion classification systems which work across domains can be developed with reasonable effort. For a productive application of emotion detection in the context of speech events we conclude that a deployed system might perform best with a speech-to-text module followed by an analysis of the text. Further, in this work, we did not explore an ensemble model or the interaction of different modalities. Thus, future work should investigate the fusion of multiple modalities in a single classifier. ### Acknowledgment
We thank Laura-Ana-Maria Bostan for discussions and data set preparations. This research has partially been funded by the German Research Council (DFG), project SEAT (KL 2869/1-1). Figure 1: The setup of the driving simulator. Table 1: Examples for triggered interactions with translations to English. (D: Driver, A: Agent, Co: Co-Driver) Table 2: Examples from the collected data set (with translation to English). E: Emotion, IT: interaction type with agent (A) and with Codriver (C). J: Joy, A: Annoyance, I: Insecurity, B: Boredom, R: Relaxation, N: No emotion. Figure8 8,419 1,419 9,179 19,017 EmoInt 2,252 1,701 1,616 5,569 ISEAR 1,095 1,096 1,094 3,285 TEC 2,782 1,534 8,132 12,448 AMMER 49 26 90 165 Figure 2: Model for Transfer Learning from Text. Grey boxes contain frozen parameters in the corresponding learning step. Figure8 66 55 59 76 EmoInt 62 48 56 76 TEC 73 55 58 76 ISEAR 70 35 59 72 AMMER 57 — — — Table 4: Confusion Matrix for Face Classification and Audio Classification (on full AMMER data) and for transfer learning from text (training set of EmoInt and test set of AMMER). Insecurity, annoyance and joy are the gold labels. Fear, anger and joy are predictions. Table 5: Performance for classification from vision, audio, and transfer learning from text (training set of EmoInt). | confusion matrices, $\text{F}_1$ score |
What is CamemBERT trained on? | ### Introduction
Pretrained word representations have a long history in Natural Language Processing (NLP), from non-neural methods BIBREF0, BIBREF1, BIBREF2 to neural word embeddings BIBREF3, BIBREF4 and to contextualised representations BIBREF5, BIBREF6. Approaches shifted more recently from using these representations as an input to task-specific architectures to replacing these architectures with large pretrained language models. These models are then fine-tuned to the task at hand with large improvements in performance over a wide range of tasks BIBREF7, BIBREF8, BIBREF9, BIBREF10. These transfer learning methods exhibit clear advantages over more traditional task-specific approaches, probably the most important being that they can be trained in an unsupervised manner. They nevertheless come with implementation challenges, namely the amount of data and computational resources needed for pretraining that can reach hundreds of gigabytes of uncompressed text and require hundreds of GPUs BIBREF11, BIBREF9. The latest transformer architecture has gone uses as much as 750GB of plain text and 1024 TPU v3 for pretraining BIBREF10. This has limited the availability of these state-of-the-art models to the English language, at least in the monolingual setting. Even though multilingual models give remarkable results, they are often larger and their results still lag behind their monolingual counterparts BIBREF12. This is particularly inconvenient as it hinders their practical use in NLP systems as well as the investigation of their language modeling capacity, something that remains to be investigated in the case of, for instance, morphologically rich languages. We take advantage of the newly available multilingual corpus OSCAR BIBREF13 and train a monolingual language model for French using the RoBERTa architecture. We pretrain the model - which we dub CamemBERT- and evaluate it in four different downstream tasks for French: part-of-speech (POS) tagging, dependency parsing, named entity recognition (NER) and natural language inference (NLI). CamemBERT improves the state of the art for most tasks over previous monolingual and multilingual approaches, which confirms the effectiveness of large pretrained language models for French. We summarise our contributions as follows: We train a monolingual BERT model on the French language using recent large-scale corpora. We evaluate our model on four downstream tasks (POS tagging, dependency parsing, NER and natural language inference (NLI)), achieving state-of-the-art results in most tasks, confirming the effectiveness of large BERT-based models for French. We release our model in a user-friendly format for popular open-source libraries so that it can serve as a strong baseline for future research and be useful for French NLP practitioners. ### Related Work ::: From non-contextual to contextual word embeddings
The first neural word vector representations were non-contextualised word embeddings, most notably word2vec BIBREF3, GloVe BIBREF4 and fastText BIBREF14, which were designed to be used as input to task-specific neural architectures. Contextualised word representations such as ELMo BIBREF5 and flair BIBREF6, improved the expressivity of word embeddings by taking context into account. They improved the performance of downstream tasks when they replaced traditional word representations. This paved the way towards larger contextualised models that replaced downstream architectures in most tasks. These approaches, trained with language modeling objectives, range from LSTM-based architectures such as ULMFiT BIBREF15 to the successful transformer-based architectures such as GPT2 BIBREF8, BERT BIBREF7, RoBERTa BIBREF9 and more recently ALBERT BIBREF16 and T5 BIBREF10. ### Related Work ::: Non-contextual word embeddings for languages other than English
Since the introduction of word2vec BIBREF3, many attempts have been made to create monolingual models for a wide range of languages. For non-contextual word embeddings, the first two attempts were by BIBREF17 and BIBREF18 who created word embeddings for a large number of languages using Wikipedia. Later BIBREF19 trained fastText word embeddings for 157 languages using Common Crawl and showed that using crawled data significantly increased the performance of the embeddings relatively to those trained only on Wikipedia. ### Related Work ::: Contextualised models for languages other than English
Following the success of large pretrained language models, they were extended to the multilingual setting with multilingual BERT , a single multilingual model for 104 different languages trained on Wikipedia data, and later XLM BIBREF12, which greatly improved unsupervised machine translation. A few monolingual models have been released: ELMo models for Japanese, Portuguese, German and Basque and BERT for Simplified and Traditional Chinese and German. However, to the best of our knowledge, no particular effort has been made toward training models for languages other than English, at a scale similar to the latest English models (e.g. RoBERTa trained on more than 100GB of data). ### CamemBERT
Our approach is based on RoBERTa BIBREF9, which replicates and improves the initial BERT by identifying key hyper-parameters for more robust performance. In this section, we describe the architecture, training objective, optimisation setup and pretraining data that was used for CamemBERT. CamemBERT differs from RoBERTa mainly with the addition of whole-word masking and the usage of SentencePiece tokenisation BIBREF20. ### CamemBERT ::: Architecture
Similar to RoBERTa and BERT, CamemBERT is a multi-layer bidirectional Transformer BIBREF21. Given the widespread usage of Transformers, we do not describe them in detail here and refer the reader to BIBREF21. CamemBERT uses the original BERT $_{\small \textsc {BASE}}$ configuration: 12 layers, 768 hidden dimensions, 12 attention heads, which amounts to 110M parameters. ### CamemBERT ::: Pretraining objective
We train our model on the Masked Language Modeling (MLM) task. Given an input text sequence composed of $N$ tokens $x_1, ..., x_N$, we select $15\%$ of tokens for possible replacement. Among those selected tokens, 80% are replaced with the special $<$mask$>$ token, 10% are left unchanged and 10% are replaced by a random token. The model is then trained to predict the initial masked tokens using cross-entropy loss. Following RoBERTa we dynamically mask tokens instead of fixing them statically for the whole dataset during preprocessing. This improves variability and makes the model more robust when training for multiple epochs. Since we segment the input sentence into subwords using SentencePiece, the input tokens to the models can be subwords. An upgraded version of BERT and BIBREF22 have shown that masking whole words instead of individual subwords leads to improved performance. Whole-word masking (WWM) makes the training task more difficult because the model has to predict a whole word instead of predicting only part of the word given the rest. As a result, we used WWM for CamemBERT by first randomly sampling 15% of the words in the sequence and then considering all subword tokens in each of these 15% words for candidate replacement. This amounts to a proportion of selected tokens that is close to the original 15%. These tokens are then either replaced by $<$mask$>$ tokens (80%), left unchanged (10%) or replaced by a random token. Subsequent work has shown that the next sentence prediction task (NSP) originally used in BERT does not improve downstream task performance BIBREF12, BIBREF9, we do not use NSP as a consequence. ### CamemBERT ::: Optimisation
Following BIBREF9, we optimise the model using Adam BIBREF23 ($\beta _1 = 0.9$, $\beta _2 = 0.98$) for 100k steps. We use large batch sizes of 8192 sequences. Each sequence contains at most 512 tokens. We enforce each sequence to only contain complete sentences. Additionally, we used the DOC-SENTENCES scenario from BIBREF9, consisting of not mixing multiple documents in the same sequence, which showed slightly better results. ### CamemBERT ::: Segmentation into subword units
We segment the input text into subword units using SentencePiece BIBREF20. SentencePiece is an extension of Byte-Pair encoding (BPE) BIBREF24 and WordPiece BIBREF25 that does not require pre-tokenisation (at the word or token level), thus removing the need for language-specific tokenisers. We use a vocabulary size of 32k subword tokens. These are learned on $10^7$ sentences sampled from the pretraining dataset. We do not use subword regularisation (i.e. sampling from multiple possible segmentations) in our implementation for simplicity. ### CamemBERT ::: Pretraining data
Pretrained language models can be significantly improved by using more data BIBREF9, BIBREF10. Therefore we used French text extracted from Common Crawl, in particular, we use OSCAR BIBREF13 a pre-classified and pre-filtered version of the November 2018 Common Craw snapshot. OSCAR is a set of monolingual corpora extracted from Common Crawl, specifically from the plain text WET format distributed by Common Crawl, which removes all HTML tags and converts all text encodings to UTF-8. OSCAR follows the same approach as BIBREF19 by using a language classification model based on the fastText linear classifier BIBREF26, BIBREF27 pretrained on Wikipedia, Tatoeba and SETimes, which supports 176 different languages. OSCAR performs a deduplication step after language classification and without introducing a specialised filtering scheme, other than only keeping paragraphs containing 100 or more UTF-8 encoded characters, making OSCAR quite close to the original Crawled data. We use the unshuffled version of the French OSCAR corpus, which amounts to 138GB of uncompressed text and 32.7B SentencePiece tokens. ### Evaluation ::: Part-of-speech tagging and dependency parsing
We fist evaluate CamemBERT on the two downstream tasks of part-of-speech (POS) tagging and dependency parsing. POS tagging is a low-level syntactic task, which consists in assigning to each word its corresponding grammatical category. Dependency parsing consists in predicting the labeled syntactic tree capturing the syntactic relations between words. We run our experiments using the Universal Dependencies (UD) paradigm and its corresponding UD POS tag set BIBREF28 and UD treebank collection version 2.2 BIBREF29, which was used for the CoNLL 2018 shared task. We perform our work on the four freely available French UD treebanks in UD v2.2: GSD, Sequoia, Spoken, and ParTUT. GSD BIBREF30 is the second-largest treebank available for French after the FTB (described in subsection SECREF25), it contains data from blogs, news articles, reviews, and Wikipedia. The Sequoia treebank BIBREF31, BIBREF32 comprises more than 3000 sentences, from the French Europarl, the regional newspaper L’Est Républicain, the French Wikipedia and documents from the European Medicines Agency. Spoken is a corpus converted automatically from the Rhapsodie treebank BIBREF33, BIBREF34 with manual corrections. It consists of 57 sound samples of spoken French with orthographic transcription and phonetic transcription aligned with sound (word boundaries, syllables, and phonemes), syntactic and prosodic annotations. Finally, ParTUT is a conversion of a multilingual parallel treebank developed at the University of Turin, and consisting of a variety of text genres, including talks, legal texts, and Wikipedia articles, among others; ParTUT data is derived from the already-existing parallel treebank Par(allel)TUT BIBREF35 . Table TABREF23 contains a summary comparing the sizes of the treebanks. We evaluate the performance of our models using the standard UPOS accuracy for POS tagging, and Unlabeled Attachment Score (UAS) and Labeled Attachment Score (LAS) for dependency parsing. We assume gold tokenisation and gold word segmentation as provided in the UD treebanks. ### Evaluation ::: Part-of-speech tagging and dependency parsing ::: Baselines
To demonstrate the value of building a dedicated version of BERT for French, we first compare CamemBERT to the multilingual cased version of BERT (designated as mBERT). We then compare our models to UDify BIBREF36. UDify is a multitask and multilingual model based on mBERT that is near state-of-the-art on all UD languages including French for both POS tagging and dependency parsing. It is relevant to compare CamemBERT to UDify on those tasks because UDify is the work that pushed the furthest the performance in fine-tuning end-to-end a BERT-based model on downstream POS tagging and dependency parsing. Finally, we compare our model to UDPipe Future BIBREF37, a model ranked 3rd in dependency parsing and 6th in POS tagging during the CoNLL 2018 shared task BIBREF38. UDPipe Future provides us a strong baseline that does not make use of any pretrained contextual embedding. We will compare to the more recent cross-lingual language model XLM BIBREF12, as well as the state-of-the-art CoNLL 2018 shared task results with predicted tokenisation and segmentation in an updated version of the paper. ### Evaluation ::: Named Entity Recognition
Named Entity Recognition (NER) is a sequence labeling task that consists in predicting which words refer to real-world objects, such as people, locations, artifacts and organisations. We use the French Treebank (FTB) BIBREF39 in its 2008 version introduced by cc-clustering:09short and with NER annotations by sagot2012annotation. The NER-annotated FTB contains more than 12k sentences and more than 350k tokens extracted from articles of the newspaper Le Monde published between 1989 and 1995. In total, it contains 11,636 entity mentions distributed among 7 different types of entities, namely: 2025 mentions of “Person”, 3761 of “Location”, 2382 of “Organisation”, 3357 of “Company”, 67 of “Product”, 15 of “POI” (Point of Interest) and 29 of “Fictional Character”. A large proportion of the entity mentions in the treebank are multi-word entities. For NER we therefore report the 3 metrics that are commonly used to evaluate models: precision, recall, and F1 score. Here precision measures the percentage of entities found by the system that are correctly tagged, recall measures the percentage of named entities present in the corpus that are found and the F1 score combines both precision and recall measures giving a general idea of a model's performance. ### Evaluation ::: Named Entity Recognition ::: Baselines
Most of the advances in NER haven been achieved on English, particularly focusing on the CoNLL 2003 BIBREF40 and the Ontonotes v5 BIBREF41, BIBREF42 English corpora. NER is a task that was traditionally tackled using Conditional Random Fields (CRF) BIBREF43 which are quite suited for NER; CRFs were later used as decoding layers for Bi-LSTM architectures BIBREF44, BIBREF45 showing considerable improvements over CRFs alone. These Bi-LSTM-CRF architectures were later enhanced with contextualised word embeddings which yet again brought major improvements to the task BIBREF5, BIBREF6. Finally, large pretrained architectures settled the current state of the art showing a small yet important improvement over previous NER-specific architectures BIBREF7, BIBREF46. In non-English NER the CoNLL 2002 shared task included NER corpora for Spanish and Dutch corpora BIBREF47 while the CoNLL 2003 included a German corpus BIBREF40. Here the recent efforts of BIBREF48 settled the state of the art for Spanish and Dutch, while BIBREF6 did it for German. In French, no extensive work has been done due to the limited availability of NER corpora. We compare our model with the strong baselines settled by BIBREF49, who trained both CRF and BiLSTM-CRF architectures on the FTB and enhanced them using heuristics and pretrained word embeddings. ### Evaluation ::: Natural Language Inference
We also evaluate our model on the Natural Language Inference (NLI) task, using the French part of the XNLI dataset BIBREF50. NLI consists in predicting whether a hypothesis sentence is entailed, neutral or contradicts a premise sentence. The XNLI dataset is the extension of the Multi-Genre NLI (MultiNLI) corpus BIBREF51 to 15 languages by translating the validation and test sets manually into each of those languages. The English training set is also machine translated for all languages. The dataset is composed of 122k train, 2490 valid and 5010 test examples. As usual, NLI performance is evaluated using accuracy. To evaluate a model on a language other than English (such as French), we consider the two following settings: TRANSLATE-TEST: The French test set is machine translated into English, and then used with an English classification model. This setting provides a reasonable, although imperfect, way to circumvent the fact that no such data set exists for French, and results in very strong baseline scores. TRANSLATE-TRAIN: The French model is fine-tuned on the machine-translated English training set and then evaluated on the French test set. This is the setting that we used for CamemBERT. ### Evaluation ::: Natural Language Inference ::: Baselines
For the TRANSLATE-TEST setting, we report results of the English RoBERTa to act as a reference. In the TRANSLATE-TRAIN setting, we report the best scores from previous literature along with ours. BiLSTM-max is the best model in the original XNLI paper, mBERT which has been reported in French in BIBREF52 and XLM (MLM+TLM) is the best-presented model from BIBREF50. ### Experiments
In this section, we measure the performance of CamemBERT by evaluating it on the four aforementioned tasks: POS tagging, dependency parsing, NER and NLI. ### Experiments ::: Experimental Setup ::: Pretraining
We use the RoBERTa implementation in the fairseq library BIBREF53. Our learning rate is warmed up for 10k steps up to a peak value of $0.0007$ instead of the original $0.0001$ given our large batch size (8192). The learning rate fades to zero with polynomial decay. We pretrain our model on 256 Nvidia V100 GPUs (32GB each) for 100k steps during 17h. ### Experiments ::: Experimental Setup ::: Fine-tuning
For each task, we append the relevant predictive layer on top of CamemBERT's Transformer architecture. Following the work done on BERT BIBREF7, for sequence tagging and sequence labeling we append a linear layer respectively to the $<$s$>$ special token and to the first subword token of each word. For dependency parsing, we plug a bi-affine graph predictor head as inspired by BIBREF54 following the work done on multilingual parsing with BERT by BIBREF36. We refer the reader to these two articles for more details on this module. We fine-tune independently CamemBERT for each task and each dataset. We optimise the model using the Adam optimiser BIBREF23 with a fixed learning rate. We run a grid search on a combination of learning rates and batch sizes. We select the best model on the validation set out of the 30 first epochs. Although this might push the performances even further, for all tasks except NLI, we don't apply any regularisation techniques such as weight decay, learning rate warm-up or discriminative fine-tuning. We show that fine-tuning CamemBERT in a straight-forward manner leads to state-of-the-art results on most tasks and outperforms the existing BERT-based models in most cases. The POS tagging, dependency parsing, and NER experiments are run using hugging face's Transformer library extended to support CamemBERT and dependency parsing BIBREF55. The NLI experiments use the fairseq library following the RoBERTa implementation. ### Experiments ::: Results ::: Part-of-Speech tagging and dependency parsing
For POS tagging and dependency parsing, we compare CamemBERT to three other near state-of-the-art models in Table TABREF32. CamemBERT outperforms UDPipe Future by a large margin for all treebanks and all metrics. Despite a much simpler optimisation process, CamemBERT beats UDify performances on all the available French treebanks. CamemBERT also demonstrates higher performances than mBERT on those tasks. We observe a larger error reduction for parsing than for tagging. For POS tagging, we observe error reductions of respectively 0.71% for GSD, 0.81% for Sequoia, 0.7% for Spoken and 0.28% for ParTUT. For parsing, we observe error reductions in LAS of 2.96% for GSD, 3.33% for Sequoia, 1.70% for Spoken and 1.65% for ParTUT. ### Experiments ::: Results ::: Natural Language Inference: XNLI
On the XNLI benchmark, CamemBERT obtains improved performance over multilingual language models on the TRANSLATE-TRAIN setting (81.2 vs. 80.2 for XLM) while using less than half the parameters (110M vs. 250M). However, its performance still lags behind models trained on the original English training set in the TRANSLATE-TEST setting, 81.2 vs. 82.91 for RoBERTa. It should be noted that CamemBERT uses far fewer parameters than RoBERTa (110M vs. 355M parameters). ### Experiments ::: Results ::: Named-Entity Recognition
For named entity recognition, our experiments show that CamemBERT achieves a slightly better precision than the traditional CRF-based SEM architectures described above in Section SECREF25 (CRF and Bi-LSTM+CRF), but shows a dramatic improvement in finding entity mentions, raising the recall score by 3.5 points. Both improvements result in a 2.36 point increase in the F1 score with respect to the best SEM architecture (BiLSTM-CRF), giving CamemBERT the state of the art for NER on the FTB. One other important finding is the results obtained by mBERT. Previous work with this model showed increased performance in NER for German, Dutch and Spanish when mBERT is used as contextualised word embedding for an NER-specific model BIBREF48, but our results suggest that the multilingual setting in which mBERT was trained is simply not enough to use it alone and fine-tune it for French NER, as it shows worse performance than even simple CRF models, suggesting that monolingual models could be better at NER. ### Experiments ::: Discussion
CamemBERT displays improved performance compared to prior work for the 4 downstream tasks considered. This confirms the hypothesis that pretrained language models can be effectively fine-tuned for various downstream tasks, as observed for English in previous work. Moreover, our results also show that dedicated monolingual models still outperform multilingual ones. We explain this point in two ways. First, the scale of data is possibly essential to the performance of CamemBERT. Indeed, we use 138GB of uncompressed text vs. 57GB for mBERT. Second, with more data comes more diversity in the pretraining distribution. Reaching state-of-the-art performances on 4 different tasks and 6 different datasets requires robust pretrained models. Our results suggest that the variability in the downstream tasks and datasets considered is handled more efficiently by a general language model than by Wikipedia-pretrained models such as mBERT. ### Conclusion
CamemBERT improves the state of the art for multiple downstream tasks in French. It is also lighter than other BERT-based approaches such as mBERT or XLM. By releasing our model, we hope that it can serve as a strong baseline for future research in French NLP, and expect our experiments to be reproduced in many other languages. We will publish an updated version in the near future where we will explore and release models trained for longer, with additional downstream tasks, baselines (e.g. XLM) and analysis, we will also train additional models with potentially cleaner corpora such as CCNet BIBREF56 for more accurate performance evaluation and more complete ablation. ### Acknowledgments
This work was partly funded by three French National grants from the Agence Nationale de la Recherche, namely projects PARSITI (ANR-16-CE33-0021), SoSweet (ANR-15-CE38-0011) and BASNUM (ANR-18-CE38-0003), as well as by the last author's chair in the PRAIRIE institute. ### Appendix ::: Impact of Whole-Word Masking
We analyze the addition of whole-word masking on the downstream performance of CamemBERT. As reported for English on other downstream tasks, whole word masking improves downstream performances for all tasks but NER as seen in Table TABREF46. NER is highly sensitive to capitalisation, prefixes, suffixes and other subword features that could be used by a model to correctly identify entity mentions. Thus the added information by learning the masking at a subword level rather than at whole-word level seems to have a detrimental effect on downstream NER results. Table 1: Sizes in Number of tokens, words and phrases of the 4 treebanks used in the evaluations of POS-tagging and dependency parsing. Table 2: Final POS and dependency parsing scores of CamemBERT and mBERT (fine-tuned in the exact same conditions as CamemBERT), UDify as reported in the original paper on 4 French treebanks (French GSD, Spoken, Sequoia and ParTUT), reported on test sets (4 averaged runs) assuming gold tokenisation. Best scores in bold, second to best underlined. Table 3: Accuracy of models for French on the XNLI test set. Best scores in bold, second to best underlined. Table 4: Results for NER on the FTB. Best scores in bold, second to best underlined. Table 5: Comparing subword and whole-word masking procedures on the validation sets of each task. Each score is an average of 4 runs with different random seeds. For POS tagging and Dependency parsing, we average the scores on the 4 treebanks.) | unshuffled version of the French OSCAR corpus |
Given the details in the article, what best describes Captain Walsh and Major Polk's relationship?
A. They had strong disdain for each other.
B. They often bantered while still being close friends.
C. They enjoyed competing with each other.
D. They liked to make jokes out of each other.
| A PLANET NAMED JOE By S. A. LOMBINO There were more Joes on Venus than you could shake a ray-gun at. Perhaps there was method in Colonel Walsh's madness—murder-madness—when he ordered Major Polk to scan the planet for a guy named Joe. [Transcriber's Note: This etext was produced from Planet Stories November 1952. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed.] Colonel Walsh had a great sense of humor. I hated his guts ever since we went through the Academy together, but he had a great sense of humor. For example, he could have chosen a Second Looie for the job on Venus. He might even have picked a Captain. But he liked me about as much as I liked him, and so he decided the job was just right for a Major. At least, that's what he told me. I stood at attention before his desk in the Patrol Station. We were somewhere in Area Two on Earth, takeoff point for any operations in Space II. The duty was fine, and I liked it a lot. Come to think of it, the most I ever did was inspect a few defective tubes every now and then. The rest was gravy, and Colonel Walsh wasn't going to let me get by with gravy. "It will be a simple assignment, Major," he said to me, peering over his fingers. He held them up in front of him like a cathedral. "Yes, sir," I said. "It will involve finding one man, a Venusian native." I wanted to say, "Then why the hell don't you send a green kid on the job? Why me?" Instead, I nodded and watched him playing with his fingers. "The man is a trader of sorts. Rather intelligent." He paused, then added, "For a native, that is." I had never liked Walsh's attitude toward natives. I hadn't liked the way he'd treated the natives on Mars ever since he'd taken over there. Which brought to mind an important point. "I always figured Venus was under the jurisdiction of Space III, sir. I thought our activities were confined to Mars." He folded his fingers like a deck of cards and dropped them on his desk as if he were waiting for me to cut. "Mmmm," he said, "yes, that's true. But this is a special job. It so happens this Venusian is the one man who can help us understand just what's happening on Mars." I tried to picture a Venusian understanding Mars and I didn't get very far. "He's had many dealings with the natives there," Walsh explained. "If anyone can tell us the reasons for the revolt, he can." If Walsh really wanted to know the reasons for the revolt, I could give them to him in one word: Walsh. I had to laugh at the way he called it "revolt." It had been going on for six months now and we'd lost at least a thousand men from Space II. Revolt. "And this man is on Venus now?" I asked for confirmation. I'd never been to Venus, being in Space II ever since I'd left the Moon run. It was just like Walsh to ship me off to a strange place. "Yes, Major," he said. "This man is on Venus." At the Academy he had called me Fred. That was before I'd reported him for sleeping on Boiler Watch. He'd goofed off on a pile of uranium that could've, and almost did, blow the barracks sky-high that night. He still thought it was my fault, as if I'd done the wrong thing by reporting him. And now, through the fouled-up machinery that exists in any military organization, he outranked me. "And the man's name, sir?" "Joe." A tight smile played on his face. "Joe what?" I asked. "Just Joe." "Just Joe?" "Yes," Walsh said. "A native, you know. They rarely go in for more than first names. But then, it should be simple to find a man with a name like Joe. Among the natives, I mean." "I don't know, sir." "A relatively simple assignment," Walsh said. "Can you tell me anything else about this man? Physical appearance? Personal habits? Anything?" Walsh seemed to consider this for a moment. "Well, physically he's like any of the other Venusians, so I can't give you much help there. He does have a peculiar habit, though." "What's that?" "He has an affinity for Terran cigarettes." I sighed. "Well, it's not very much to go on." "You'll find him," Walsh said, grinning. "I'm sure of it." The trip to Venus came off without a hitch. I did a lot of thinking on that trip. I thought about Mars and the revolt there. And I thought about Colonel Leonard Walsh and how he was supposed to be quelling that revolt. Ever since Walsh had taken command, ever since he'd started pushing the natives around, there'd been trouble. It was almost as if the whole damned planet had blown up in our faces the moment he took over. Swell guy, Walsh. Venus was hotter than I'd expected it to be. Much too hot for the tunic I was wearing. It smelled, too. A funny smell I couldn't place. Like a mixture of old shoe and after-shave. There were plants everywhere I looked. Big plants and small ones, some blooming with flowers I'd never seen before, and some as bare as cactus. I recognized a blue figure as one of the natives the pilot had told me about. He was tall, looking almost human except that everything about him was elongated. His features, his muscles, everything seemed to have been stretched like a rubber band. I kept expecting him to pop back to normal. Instead, he flashed a double row of brilliant teeth at me. I wondered if he spoke English. "Hey, boy," I called. He ambled over with long-legged strides that closed the distance between us in seconds. "Call me Joe," he said. I dropped my bags and stared at him. Maybe this was going to be a simple assignment after all. "I sure am glad to see you, Joe," I said. "Same here, Toots," he answered. "The guys back in Space II are searching high and low for you," I told him. "You've got the wrong number," he said, and I was a little surprised at his use of Terran idiom. "You are Joe, aren't you? Joe the trader?" "I'm Joe, all right," he said. "Only thing I ever traded, though, was a pocketknife. Got a set of keys for it." "Oh," I said, my voice conveying my disappointment. I sighed and began wondering just how I should go about contacting the Joe I was looking for. My orders said I was to report to Captain Bransten immediately upon arrival. I figured the hell with Captain Bransten. I outranked him anyway, and there wasn't much he could do if I decided to stop for a drink first. "Where's the Officer's Club?" I asked the Venusian. "Are you buying information or are you just curious?" "Can you take me there?" I asked. "Sure thing, Toots." He picked up my bags and started walking up a heavily overgrown path. We'd probably walked for about ten minutes when he dropped my bags and said, "There it is." The Officer's Club was a plasteel hut with window shields that protected it from the heat of the sun. It didn't look too comfortable but I really wanted that drink. I reached into my tunic and slipped the native thirty solars. He stared at the credits curiously and then shrugged his shoulders. "Oh well, you're new here. We'll let it go." He took off then, while I stared after him, wondering just what he'd meant. Had I tipped him too little? I shrugged and looked over at the Officer's Club. From the outside it looked as hot as hell. On the inside it was about two degrees short of that mark. I began to curse Walsh for taking me away from my nice soft job in Space II. There wasn't much inside the club. A few tables and chairs, a dart game and a bar. Behind the bar a tall Venusian lounged. I walked over and asked, "What are you serving, pal?" "Call me Joe," he answered. He caught me off balance. "What?" "Joe," he said again. A faint glimmer of understanding began to penetrate my thick skull. "You wouldn't happen to be Joe the trader? The guy who knows all about Mars, would you?" "I never left home," he said simply. "What are you drinking?" That rat! That dirty, filthy, stinking, unprincipled.... But then, it should be simple to find a man with a name like Joe. Among the natives, I mean. Sure. Oh sure. Real simple. Walsh was about the lowest, most contemptible.... "What are you drinking, pal?" the Venusian asked again. "Skip it," I said. "How do I get to the captain's shack?" "Follow your nose, pal. Can't miss it." I started to pick up my bag as another Venusian entered. He waved at the bartender. "Hello, Joe," he said. "How's it going?" "Not so hot, Joe," the bartender replied. I listened in fascination. Joe, Joe, Joe. So this was Walsh's idea of a great gag. Very funny. Very.... "You Major Polk, sweetheart?" the Venusian who'd just come in asked. "Yes," I said, still thinking of Colonel Walsh. "You better get your butt over to the captain's shack," he said. "He's about ready to post you as overdue." "Sure," I said wearily. "Will you take my bags, please?" "Roger," he answered. He picked up the bags and nodded at the bar. "So long, Joe," he said to the bartender. "See you, Joe," the bartender called back. Captain Bransten was a mousey, unimpressive sort of man. He was wearing a tropical tunic, but he still resembled a wilted lily more than he did an officer. "Have a seat, Major," he offered. He reached for a cigarette box on the desk and extended it to me. He coughed in embarrassment when he saw it was empty. Quickly, he pressed a button on his desk and the door popped open. A tall, blue Venusian stepped lithely into the room. "Sir?" the Venusian asked. "We're out of cigarettes, Joe," the Captain said. "Will you get us some, please?" "Sure thing," the Venusian answered. He smiled broadly and closed the door behind him. Another Joe , I thought. Another damned Joe. "They steal them," Captain Bransten said abruptly. "Steal what?" I asked. "Cigarettes. I sometimes think the cigarette is one of the few things they like about Terran culture." So Walsh had taken care of that angle too. He does have a peculiar habit, though. He has an affinity for Terran cigarettes. Cigarettes was the tip I should have given; not solars. "All right," I said, "suppose we start at the beginning." Captain Bransten opened his eyes wide. "Sir?" he asked. "What's with all this Joe business? It may be a very original name but I think its popularity here is a little outstanding." Captain Bransten began to chuckle softly. I personally didn't think it was so funny. I tossed him my withering Superior Officer's gaze and waited for his explanation. "I hadn't realized this was your first time on Venus," he said. "Is there a local hero named Joe?" I asked. "No, no, nothing like that," he assured me. "It's a simple culture, you know. Not nearly as developed as Mars." "I can see that," I said bitingly. "And the natives are only now becoming acquainted with Terran culture. Lots of enlisted men, you know." I began to get the idea. And I began to appreciate Walsh's doubtful ancestry more keenly. "It's impossible to tell exactly where it all started, of course," Bransten was saying. I was beginning to get angry. Very angry. I was thinking of Walsh sitting back in a nice cozy foam chair back on Earth. "Get to the point, Captain!" I barked. "Easy, sir," Bransten said, turning pale. I could see that the Captain wasn't used to entertaining Majors. "The enlisted men. You know how they are. They'll ask a native to do something and they'll call him Joe. 'Hey, Joe, give me a hand with this.' Or 'Listen, Joe, how'd you like to earn some cigarettes?' Do you follow?" "I follow, all right," I said bitterly. "Well," Bransten went on, "that sort of thing mushrooms. The natives are a simple, almost childish people. It appealed to them—the Joe business, I mean. Now they're all Joe. They like it. That and the cigarettes." He cleared his throat and looked at me apologetically as if he were personally responsible for Venusian culture. In fact, he looked as if he were responsible for having put Venus in the heavens in the first place. "Do you understand, Major? Just a case of extended idiom, that's all." Just a case of extended idiot , I thought. An idiot on a wild goose chase a hell of a long way from home. "I understand perfectly," I snapped. "Where are my quarters?" Bransten asked a Venusian named Joe to show me my quarters, reminding me that chow was at thirteen hundred. As I was leaving, the first Venusian came back with the cigarettes Bransten had ordered. I could tell by the look on his face that he probably had half a carton stuffed into his pockets. I shrugged and went to change into a tropical tunic. I called Earth right after chow. The Captain assured me that this sort of thing was definitely against regulations, but he submitted when I twinkled my little gold leaf under his nose. Walsh's face appeared on the screen. He was smiling, looking like a fat pussy cat. "What is it, Major?" he asked. "This man Joe," I said. "Can you give me any more on him?" Walsh's grin grew wider. "Why, Major," he said, "you're not having any difficulties, are you?" "None at all," I snapped back. "I just thought I'd be able to find him a lot sooner if...." "Take your time, Major," Walsh beamed. "There's no rush at all." "I thought...." "I'm sure you can do the job," Walsh cut in. "I wouldn't have sent you otherwise." Hell, I was through kidding around. "Look...." "He's somewhere in the jungle, you know," Walsh said. I wanted to ram my fist into the screen, right smack up against those big white teeth. Instead, I cut off the transmission and watched the surprised look on his face as his screen went blank millions of miles away. He blinked at the screen, trying to realize I'd deliberately hung up on him. "Polk!" he shouted, "can you hear me?" I smiled, saw the twisted hatred on his features, and then the screen on my end went blank, too. He's somewhere in the jungle, you know. I thanked Captain Bransten for his hospitality and went back to my quarters. As I saw it, there were two courses for me to follow. One: I could say the hell with Walsh and Venus. That would mean hopping the next ship back to Earth. It would also mean disobeying the direct order of a superior officer. It might mean demotion, and it might mean getting bounced out of the Service altogether. Two: I could assume there really was a guy name Joe somewhere in that jungle, a Joe separate and apart from the other Joes on this planet, a trader Joe who knew the Martians well. I could always admit failure, of course, and return empty handed. Mission not accomplished. Or, I might really find a guy who was trader Joe. I made my decision quickly. I wanted to stay in the Service, and besides Walsh may have been on the level for the first time in his life. Maybe there was a Joe here who could help us on Mars. If there was I'd try to find him. It was still a hell of a trick though. I cursed Walsh again and pushed the buzzer near my bed. A tall Venusian stepped into the room. "Joe?" I asked, just to be sure. "Who else, boss?" he answered. "I'm trying to locate someone," I said. "I'll need a guide to take me into the jungle. Can you get me one?" "It'll cost you, boss," the Venusian said. "How much?" "Two cartons of cigarettes at least." "Who's the guide?" I asked. "How's the price sound?" "Fine, fine," I said impatiently. And the Captain had said they were almost a childish people! "His name is Joe," the Venusian told me. "Best damn guide on the planet. Take you anywhere you want to go, do anything you want to do. Courageous. Doesn't know the meaning of fear. I've known him to...." "Skip it," I said, cutting the promotion short. "Tell him to show up around fifteen hundred with a complete list of what we'll need." The Venusian started to leave. "And Joe," I said, stopping him at the door, "I hope you're not overlooking your commission on the deal." His face broke into a wide grin. "No danger of that, boss," he said. When he was gone I began figuring out a plan of action. Obviously, I'd just have to traipse through the jungle looking for a guy named Joe on a planet where everyone was named Joe. Everybody, at least, but the Captain, the small garrison attached to the Station, and me. I began wondering why Walsh had gone to so much trouble to get rid of me. The job, as I saw it, would take a hell of a long time. It seemed like a silly thing to do, just to get even with a guy for something that had happened years ago. He surely must have realized that I'd be back again, sooner or later. Maybe he had another little junket all set for me. Or maybe he didn't expect me to come back. The thought hadn't occurred to me before this, and I began to consider it seriously. Walsh was no good, rotten clear through. He was failing at the job of keeping Mars in hand, and he probably realized that a few more mistakes on his part would mean the end of his career with Space II. I chuckled as I thought of him isolated in some God-forsaken place like Space V or Space VII. This probably bothered him a lot, too. But what probably bothered him more was the fact that I was next in command. If he were transferred, I'd be in charge of Space II, and I could understand how much that would appeal to Walsh. I tried to figure the thing out sensibly, tried to weigh his good points against his bad. But it all came back to the same thing. A guy who would deliberately go to sleep on Boiler Watch with a ton of uranium ready to blast a barracks to smithereens if it wasn't watched, would deliberately do just about anything. Sending me off on a wild goose chase after a character named Joe may have been a gag. But it may have been something a little grimmer than a gag, and I made up my mind to be extremely careful from here on in. The guide arrived at fifteen hundred on the dot. He was tall, elongated, looked almost like all the other Venusians I'd seen so far. "I understand you need a Grade A guide, sir," he said. "Are you familiar with the jungle?" I asked him. "Born and raised there, sir. Know it like the back of my hand." "Has Joe told you what the payment will be?" "Yes, sir. A carton and a half of cigarettes." I thought about Joe deducting his commission and smiled. "When can we leave?" "Right away, sir. We won't need much really. I've made a list of supplies and I can get them in less than an hour. I suggest you wear light clothing, boots, and a hat." "Will I need a weapon?" He looked at me, his eyes faintly amused. "Why, what for, sir?" "Never mind," I said. "What's your name, by the way?" He lifted his eyebrows, and his eyes widened in his narrow face. He was definitely surprised. "Joe," he said. "Didn't you know?" When we'd been out for a while I discovered why Joe had suggested the boots and the hat. The undergrowth was often sharp and jagged and it would have sliced my legs to ribbons were they not protected by the high boots. The hat kept the strong sun off my head. Joe was an excellent guide and a pleasant companion. He seemed to be enjoying a great romp, seemed to love the jungle and take a secret pleasure in the work he was doing. There were times when I couldn't see three feet ahead of me. He'd stand stock still for a few minutes, his head barely moving, his eyes darting from one plant to another. Then he'd say, "This way," and take off into what looked like more impenetrable jungle invariably to find a little path leading directly to another village. Each village was the same. The natives would come running out of their huts, tall and blue, shouting, "Cigarettes, Joe? Cigarettes?" It took me a while to realize they were addressing me and not my guide. Everybody was Joe. It was one beautiful, happy, joyous round of stinking, hot jungle. And I wasn't getting any nearer my man. Nor had I any idea how I was supposed to find him. I began to feel pretty low about the whole affair. Joe, on the other hand, enjoyed every moment of the trip. In each village he greeted the natives cheerfully, told them stories, swapped gossip and jokes. And when it was time to leave, he would say goodbye to all his friends and we would plunge into the twisted foliage again. His spirits were always high and he never failed to say the right thing that would give a momentary lift to my own depressed state of mind. He would talk for hours on end as we hacked our way through the jungle. "I like Venus," he said once. "I would never leave it." "Have you ever been to Earth?" I asked. "No," Joe replied. "I like Terrans too, you understand. They are good for Venus. And they are fun." "Fun?" I asked, thinking of a particular species of Terran: species Leonard Walsh. "Yes, yes," he said wholeheartedly. "They joke and they laugh and ... well, you know." "I suppose so," I admitted. Joe smiled secretly, and we pushed on. I began to find, more and more, that I had started to talk freely to Joe. In the beginning he had been just my guide. There had been the strained relationship of employer and employee. But as the days lengthened into weeks, the formal atmosphere began to crumble. I found myself telling him all about Earth, about the people there, about my decision to attend the Academy, the rigid tests, the grind, even the Moon run. Joe was a good listener, nodding sympathetically, finding experiences in his own life to parallel my own. And as our relationship progressed from a casual one to a definitely friendly one, Joe seemed more enthusiastic than ever to keep up our grinding pace to find what we were looking for. Once we stopped in a clearing to rest. Joe lounged on the matted greenery, his long body stretched out in front of him, the knife gleaming in his belt. I'd seen him slash his way through thick, tangled vines with that knife, his long, muscular arms powerfully slicing through them like strips of silk. "How far are we from the Station?" I asked. "Three or four Earth weeks," he replied. I sighed wearily. "Where do we go from here?" "There are more villages," he said. "We'll never find him." "Possibly," Joe mused, the smile creeping over his face again. "A wild goose chase. A fool's errand." "We'd better get started," Joe said simply. I got to my feet and we started the march again. Joe was still fresh, a brilliant contrast to me, weary and dejected. Somehow, I had the same feeling I'd had a long time ago on my sixteenth birthday. One of my friends had taken me all over the city, finally dropping me off at my own house where the whole gang was gathered for a surprise party. Joe reminded me of that friend. "There's a village ahead," he said, and the grin on his face was large now, his eyes shining. Something was missing here. Natives. There were no natives rushing out to greet us. No cries of "Cigarettes? Cigarettes?" I caught up with Joe. "What's the story?" I whispered. He shrugged knowingly and continued walking. And then I saw the ship, nose pointing into space, catching the rays of the sun like a great silver bullet. "What...?" I started. "It's all right," Joe said, smiling. The ship looked vaguely familiar. I noticed the crest of Space II near the nose, and a lot of things became clear then. I also saw Walsh standing near one of the huts, a stun gun in his hand. "Hello, Major," he called, almost cheerfully. The gun didn't look cheerful, though. It was pointed at my head. "Fancy meeting you here, Colonel," I said, trying to match his joviality. Somehow it didn't quite come off. Joe was walking beside me, waving at the colonel, beaming all over with happiness. "I see you found your man," Walsh said. I turned rapidly. Joe nodded and kept grinning, a grin that told me he was getting a big kick out of all this. Like a kid playing a game. I faced Walsh again. "Okay, what's it all about, pal?" "Colonel," Walsh corrected me. "You mustn't forget to say Colonel, Major ." He emphasized my rank, and he said it with a sort of ruthless finality. I waited. I could see he was just busting to tell me how clever he'd been. Besides, there wasn't much I could do but wait. Not with Walsh pointing the stun gun at my middle. "We've come a long way since the Academy, haven't we, Major?" "If you mean in miles," I said, looking around at the plants, "we sure have." Walsh grinned a little. "Always the wit," he said drily. And then the smile faded from his lips and his eyes took on a hard lustre. "I'm going to kill you, you know." He said it as if he were saying, "I think it'll rain tomorrow." Joe almost clapped his hands together with glee. He was really enjoying this. Another of those funny Terran games. "You gave me a powerful handicap to overcome," Walsh said. "I suppose I should thank you, really." "You're welcome," I said. "It wasn't easy living down the disgrace you caused me." "It was your own damn fault," I said. "You knew what you were doing when you decided to cork off." Beside me, Joe chuckled a little, enjoying the game immensely. "You didn't have to report me," Walsh said. "No? Maybe I should have forgotten all about it? Maybe I should have nudged you and served you orange juice? So you could do it again sometime and maybe blow up the whole damn Academy!" Walsh was silent for a long time. When he spoke his voice was barely audible. The heat was oppressive, as if it were concentrated on this little spot in the jungle, focusing all its penetration on a small, unimportant drama. I could hear Joe breathing beside me. "I'm on my way out," Walsh rasped. "Finished, do you understand?" "Good," I said. And I meant it. "This Mars thing. A terrible fix. Terrible." Beside me, a slight frown crossed Joe's face. Apparently he couldn't understand the seriousness of our voices. What had happened to the game, the fun? "You brought the Mars business on yourself," I told Walsh. "There was never any trouble before you took command." "The natives," he practically shouted. "They ... they...." Joe caught his breath sharply, and I wondered what Walsh was going to say about the natives. Apparently he'd realized that Joe was a native. Or maybe Joe's knife had something to do with it. "What about the natives?" I asked. "Nothing," Walsh said. "Nothing." He was silent for a while. "A man of my calibre," he said then, his face grim. "Dealing with savages." He caught himself again and threw a hasty glance at Joe. The perplexed frown had grown heavier on Joe's face. He looked at the colonel in puzzlement. | A. They had strong disdain for each other. |
How is this approach used to detect incorrect facts? | ### Introduction
Knowledge graphs (KG) play a critical role in many real-world applications such as search, structured data management, recommendations, and question answering. Since KGs often suffer from incompleteness and noise in their facts (links), a number of recent techniques have proposed models that embed each entity and relation into a vector space, and use these embeddings to predict facts. These dense representation models for link prediction include tensor factorization BIBREF0 , BIBREF1 , BIBREF2 , algebraic operations BIBREF3 , BIBREF4 , BIBREF5 , multiple embeddings BIBREF6 , BIBREF7 , BIBREF8 , BIBREF9 , and complex neural models BIBREF10 , BIBREF11 . However, there are only a few studies BIBREF12 , BIBREF13 that investigate the quality of the different KG models. There is a need to go beyond just the accuracy on link prediction, and instead focus on whether these representations are robust and stable, and what facts they make use of for their predictions. In this paper, our goal is to design approaches that minimally change the graph structure such that the prediction of a target fact changes the most after the embeddings are relearned, which we collectively call Completion Robustness and Interpretability via Adversarial Graph Edits (). First, we consider perturbations that red!50!blackremove a neighboring link for the target fact, thus identifying the most influential related fact, providing an explanation for the model's prediction. As an example, consider the excerpt from a KG in Figure 1 with two observed facts, and a target predicted fact that Princes Henriette is the parent of Violante Bavaria. Our proposed graph perturbation, shown in Figure 1 , identifies the existing fact that Ferdinal Maria is the father of Violante Bavaria as the one when removed and model retrained, will change the prediction of Princes Henriette's child. We also study attacks that green!50!blackadd a new, fake fact into the KG to evaluate the robustness and sensitivity of link prediction models to small additions to the graph. An example attack for the original graph in Figure 1 , is depicted in Figure 1 . Such perturbations to the the training data are from a family of adversarial modifications that have been applied to other machine learning tasks, known as poisoning BIBREF14 , BIBREF15 , BIBREF16 , BIBREF17 . Since the setting is quite different from traditional adversarial attacks, search for link prediction adversaries brings up unique challenges. To find these minimal changes for a target link, we need to identify the fact that, when added into or removed from the graph, will have the biggest impact on the predicted score of the target fact. Unfortunately, computing this change in the score is expensive since it involves retraining the model to recompute the embeddings. We propose an efficient estimate of this score change by approximating the change in the embeddings using Taylor expansion. The other challenge in identifying adversarial modifications for link prediction, especially when considering addition of fake facts, is the combinatorial search space over possible facts, which is intractable to enumerate. We introduce an inverter of the original embedding model, to decode the embeddings to their corresponding graph components, making the search of facts tractable by performing efficient gradient-based continuous optimization. We evaluate our proposed methods through following experiments. First, on relatively small KGs, we show that our approximations are accurate compared to the true change in the score. Second, we show that our additive attacks can effectively reduce the performance of state of the art models BIBREF2 , BIBREF10 up to $27.3\%$ and $50.7\%$ in Hits@1 for two large KGs: WN18 and YAGO3-10. We also explore the utility of adversarial modifications in explaining the model predictions by presenting rule-like descriptions of the most influential neighbors. Finally, we use adversaries to detect errors in the KG, obtaining up to $55\%$ accuracy in detecting errors. ### Background and Notation
In this section, we briefly introduce some notations, and existing relational embedding approaches that model knowledge graph completion using dense vectors. In KGs, facts are represented using triples of subject, relation, and object, $\langle s, r, o\rangle $ , where $s,o\in \xi $ , the set of entities, and $r\in $ , the set of relations. To model the KG, a scoring function $\psi :\xi \times \times \xi \rightarrow $ is learned to evaluate whether any given fact is true. In this work, we focus on multiplicative models of link prediction, specifically DistMult BIBREF2 because of its simplicity and popularity, and ConvE BIBREF10 because of its high accuracy. We can represent the scoring function of such methods as $\psi (s,r,o) = , ) \cdot $ , where $,,\in ^d$ are embeddings of the subject, relation, and object respectively. In DistMult, $, ) = \odot $ , where $\odot $ is element-wise multiplication operator. Similarly, in ConvE, $, )$ is computed by a convolution on the concatenation of $$ and $s,o\in \xi $0 . We use the same setup as BIBREF10 for training, i.e., incorporate binary cross-entropy loss over the triple scores. In particular, for subject-relation pairs $(s,r)$ in the training data $G$ , we use binary $y^{s,r}_o$ to represent negative and positive facts. Using the model's probability of truth as $\sigma (\psi (s,r,o))$ for $\langle s,r,o\rangle $ , the loss is defined as: (G) = (s,r)o ys,ro(((s,r,o))) + (1-ys,ro)(1 - ((s,r,o))). Gradient descent is used to learn the embeddings $,,$ , and the parameters of $, if any.
$ ### Completion Robustness and Interpretability via Adversarial Graph Edits ()
For adversarial modifications on KGs, we first define the space of possible modifications. For a target triple $\langle s, r, o\rangle $ , we constrain the possible triples that we can remove (or inject) to be in the form of $\langle s^{\prime }, r^{\prime }, o\rangle $ i.e $s^{\prime }$ and $r^{\prime }$ may be different from the target, but the object is not. We analyze other forms of modifications such as $\langle s, r^{\prime }, o^{\prime }\rangle $ and $\langle s, r^{\prime }, o\rangle $ in appendices "Modifications of the Form 〈s,r ' ,o ' 〉\langle s, r^{\prime }, o^{\prime } \rangle " and "Modifications of the Form 〈s,r ' ,o〉\langle s, r^{\prime }, o \rangle " , and leave empirical evaluation of these modifications for future work. ### Removing a fact ()
For explaining a target prediction, we are interested in identifying the observed fact that has the most influence (according to the model) on the prediction. We define influence of an observed fact on the prediction as the change in the prediction score if the observed fact was not present when the embeddings were learned. Previous work have used this concept of influence similarly for several different tasks BIBREF19 , BIBREF20 . Formally, for the target triple ${s,r,o}$ and observed graph $G$ , we want to identify a neighboring triple ${s^{\prime },r^{\prime },o}\in G$ such that the score $\psi (s,r,o)$ when trained on $G$ and the score $\overline{\psi }(s,r,o)$ when trained on $G-\lbrace {s^{\prime },r^{\prime },o}\rbrace $ are maximally different, i.e. *argmax(s', r')Nei(o) (s',r')(s,r,o) where $\Delta _{(s^{\prime },r^{\prime })}(s,r,o)=\psi (s, r, o)-\overline{\psi }(s,r,o)$ , and $\text{Nei}(o)=\lbrace (s^{\prime },r^{\prime })|\langle s^{\prime },r^{\prime },o \rangle \in G \rbrace $ . ### Adding a new fact ()
We are also interested in investigating the robustness of models, i.e., how sensitive are the predictions to small additions to the knowledge graph. Specifically, for a target prediction ${s,r,o}$ , we are interested in identifying a single fake fact ${s^{\prime },r^{\prime },o}$ that, when added to the knowledge graph $G$ , changes the prediction score $\psi (s,r,o)$ the most. Using $\overline{\psi }(s,r,o)$ as the score after training on $G\cup \lbrace {s^{\prime },r^{\prime },o}\rbrace $ , we define the adversary as: *argmax(s', r') (s',r')(s,r,o) where $\Delta _{(s^{\prime },r^{\prime })}(s,r,o)=\psi (s, r, o)-\overline{\psi }(s,r,o)$ . The search here is over any possible $s^{\prime }\in \xi $ , which is often in the millions for most real-world KGs, and $r^{\prime }\in $ . We also identify adversaries that increase the prediction score for specific false triple, i.e., for a target fake fact ${s,r,o}$ , the adversary is ${s^{\prime },r^{\prime },o}$0 , where ${s^{\prime },r^{\prime },o}$1 is defined as before. ### Challenges
There are a number of crucial challenges when conducting such adversarial attack on KGs. First, evaluating the effect of changing the KG on the score of the target fact ( $\overline{\psi }(s,r,o)$ ) is expensive since we need to update the embeddings by retraining the model on the new graph; a very time-consuming process that is at least linear in the size of $G$ . Second, since there are many candidate facts that can be added to the knowledge graph, identifying the most promising adversary through search-based methods is also expensive. Specifically, the search size for unobserved facts is $|\xi | \times ||$ , which, for example in YAGO3-10 KG, can be as many as $4.5 M$ possible facts for a single target prediction. ### Efficiently Identifying the Modification
In this section, we propose algorithms to address mentioned challenges by (1) approximating the effect of changing the graph on a target prediction, and (2) using continuous optimization for the discrete search over potential modifications. ### First-order Approximation of Influence
We first study the addition of a fact to the graph, and then extend it to cover removal as well. To capture the effect of an adversarial modification on the score of a target triple, we need to study the effect of the change on the vector representations of the target triple. We use $$ , $$ , and $$ to denote the embeddings of $s,r,o$ at the solution of $\operatornamewithlimits{argmin} (G)$ , and when considering the adversarial triple $\langle s^{\prime }, r^{\prime }, o \rangle $ , we use $$ , $$ , and $$ for the new embeddings of $s,r,o$ , respectively. Thus $$0 is a solution to $$1 , which can also be written as $$2 . Similarly, $$3 s', r', o $$4 $$5 $$6 $$7 o $$8 $$9 $$0 $$1 $$2 $$3 O(n3) $$4 $$5 $$6 (s,r,o)-(s, r, o) $$7 - $$8 s, r = ,) $$9 - $s,r,o$0 (G)= (G)+(s', r', o ) $s,r,o$1 $s,r,o$2 s', r' = ',') $s,r,o$3 = ((s',r',o)) $s,r,o$4 eo (G)=0 $s,r,o$5 eo (G) $s,r,o$6 Ho $s,r,o$7 dd $s,r,o$8 o $s,r,o$9 $\operatornamewithlimits{argmin} (G)$0 - $\operatornamewithlimits{argmin} (G)$1 -= $\operatornamewithlimits{argmin} (G)$2 Ho $\operatornamewithlimits{argmin} (G)$3 Ho + (1-) s',r's',r' $\operatornamewithlimits{argmin} (G)$4 Ho $\operatornamewithlimits{argmin} (G)$5 dd $\operatornamewithlimits{argmin} (G)$6 d $\operatornamewithlimits{argmin} (G)$7 s,r,s',r'd $\operatornamewithlimits{argmin} (G)$8 s, r, o $\operatornamewithlimits{argmin} (G)$9 s', r', o $\langle s^{\prime }, r^{\prime }, o \rangle $0 $\langle s^{\prime }, r^{\prime }, o \rangle $1 $\langle s^{\prime }, r^{\prime }, o \rangle $2 ### Continuous Optimization for Search
Using the approximations provided in the previous section, Eq. () and (), we can use brute force enumeration to find the adversary $\langle s^{\prime }, r^{\prime }, o \rangle $ . This approach is feasible when removing an observed triple since the search space of such modifications is usually small; it is the number of observed facts that share the object with the target. On the other hand, finding the most influential unobserved fact to add requires search over a much larger space of all possible unobserved facts (that share the object). Instead, we identify the most influential unobserved fact $\langle s^{\prime }, r^{\prime }, o \rangle $ by using a gradient-based algorithm on vector $_{s^{\prime },r^{\prime }}$ in the embedding space (reminder, $_{s^{\prime },r^{\prime }}=^{\prime },^{\prime })$ ), solving the following continuous optimization problem in $^d$ : *argmaxs', r' (s',r')(s,r,o). After identifying the optimal $_{s^{\prime }, r^{\prime }}$ , we still need to generate the pair $(s^{\prime },r^{\prime })$ . We design a network, shown in Figure 2 , that maps the vector $_{s^{\prime },r^{\prime }}$ to the entity-relation space, i.e., translating it into $(s^{\prime },r^{\prime })$ . In particular, we train an auto-encoder where the encoder is fixed to receive the $s$ and $\langle s^{\prime }, r^{\prime }, o \rangle $0 as one-hot inputs, and calculates $\langle s^{\prime }, r^{\prime }, o \rangle $1 in the same way as the DistMult and ConvE encoders respectively (using trained embeddings). The decoder is trained to take $\langle s^{\prime }, r^{\prime }, o \rangle $2 as input and produce $\langle s^{\prime }, r^{\prime }, o \rangle $3 and $\langle s^{\prime }, r^{\prime }, o \rangle $4 , essentially inverting $\langle s^{\prime }, r^{\prime }, o \rangle $5 s, r $\langle s^{\prime }, r^{\prime }, o \rangle $6 s $\langle s^{\prime }, r^{\prime }, o \rangle $7 r $\langle s^{\prime }, r^{\prime }, o \rangle $8 s, r $\langle s^{\prime }, r^{\prime }, o \rangle $9 We evaluate the performance of our inverter networks (one for each model/dataset) on correctly recovering the pairs of subject and relation from the test set of our benchmarks, given the $_{s^{\prime },r^{\prime }}$0 . The accuracy of recovered pairs (and of each argument) is given in Table 1 . As shown, our networks achieve a very high accuracy, demonstrating their ability to invert vectors $_{s^{\prime },r^{\prime }}$1 to $_{s^{\prime },r^{\prime }}$2 pairs. ### Experiments
We evaluate by ( "Influence Function vs " ) comparing estimate with the actual effect of the attacks, ( "Robustness of Link Prediction Models" ) studying the effect of adversarial attacks on evaluation metrics, ( "Interpretability of Models" ) exploring its application to the interpretability of KG representations, and ( "Finding Errors in Knowledge Graphs" ) detecting incorrect triples. ### Influence Function vs
To evaluate the quality of our approximations and compare with influence function (IF), we conduct leave one out experiments. In this setup, we take all the neighbors of a random target triple as candidate modifications, remove them one at a time, retrain the model each time, and compute the exact change in the score of the target triple. We can use the magnitude of this change in score to rank the candidate triples, and compare this exact ranking with ranking as predicted by: , influence function with and without Hessian matrix, and the original model score (with the intuition that facts that the model is most confident of will have the largest impact when removed). Similarly, we evaluate by considering 200 random triples that share the object entity with the target sample as candidates, and rank them as above. The average results of Spearman's $\rho $ and Kendall's $\tau $ rank correlation coefficients over 10 random target samples is provided in Table 3 . performs comparably to the influence function, confirming that our approximation is accurate. Influence function is slightly more accurate because they use the complete Hessian matrix over all the parameters, while we only approximate the change by calculating the Hessian over $$ . The effect of this difference on scalability is dramatic, constraining IF to very small graphs and small embedding dimensionality ( $d\le 10$ ) before we run out of memory. In Figure 3 , we show the time to compute a single adversary by IF compared to , as we steadily grow the number of entities (randomly chosen subgraphs), averaged over 10 random triples. As it shows, is mostly unaffected by the number of entities while IF increases quadratically. Considering that real-world KGs have tens of thousands of times more entities, making IF unfeasible for them. ### Robustness of Link Prediction Models
Now we evaluate the effectiveness of to successfully attack link prediction by adding false facts. The goal here is to identify the attacks for triples in the test data, and measuring their effect on MRR and Hits@ metrics (ranking evaluations) after conducting the attack and retraining the model. Since this is the first work on adversarial attacks for link prediction, we introduce several baselines to compare against our method. For finding the adversarial fact to add for the target triple $\langle s, r, o \rangle $ , we consider two baselines: 1) choosing a random fake fact $\langle s^{\prime }, r^{\prime }, o \rangle $ (Random Attack); 2) finding $(s^{\prime }, r^{\prime })$ by first calculating $, )$ and then feeding $-, )$ to the decoder of the inverter function (Opposite Attack). In addition to , we introduce two other alternatives of our method: (1) , that uses to increase the score of fake fact over a test triple, i.e., we find the fake fact the model ranks second after the test triple, and identify the adversary for them, and (2) that selects between and attacks based on which has a higher estimated change in score. All-Test The result of the attack on all test facts as targets is provided in the Table 4 . outperforms the baselines, demonstrating its ability to effectively attack the KG representations. It seems DistMult is more robust against random attacks, while ConvE is more robust against designed attacks. is more effective than since changing the score of a fake fact is easier than of actual facts; there is no existing evidence to support fake facts. We also see that YAGO3-10 models are more robust than those for WN18. Looking at sample attacks (provided in Appendix "Sample Adversarial Attacks" ), mostly tries to change the type of the target object by associating it with a subject and a relation for a different entity type. Uncertain-Test To better understand the effect of attacks, we consider a subset of test triples that 1) the model predicts correctly, 2) difference between their scores and the negative sample with the highest score is minimum. This “Uncertain-Test” subset contains 100 triples from each of the original test sets, and we provide results of attacks on this data in Table 4 . The attacks are much more effective in this scenario, causing a considerable drop in the metrics. Further, in addition to significantly outperforming other baselines, they indicate that ConvE's confidence is much more robust. Relation Breakdown We perform additional analysis on the YAGO3-10 dataset to gain a deeper understanding of the performance of our model. As shown in Figure 4 , both DistMult and ConvE provide a more robust representation for isAffiliatedTo and isConnectedTo relations, demonstrating the confidence of models in identifying them. Moreover, the affects DistMult more in playsFor and isMarriedTo relations while affecting ConvE more in isConnectedTo relations. Examples Sample adversarial attacks are provided in Table 5 . attacks mostly try to change the type of the target triple's object by associating it with a subject and a relation that require a different entity types. ### Interpretability of Models
To be able to understand and interpret why a link is predicted using the opaque, dense embeddings, we need to find out which part of the graph was most influential on the prediction. To provide such explanations for each predictions, we identify the most influential fact using . Instead of focusing on individual predictions, we aggregate the explanations over the whole dataset for each relation using a simple rule extraction technique: we find simple patterns on subgraphs that surround the target triple and the removed fact from , and appear more than $90\%$ of the time. We only focus on extracting length-2 horn rules, i.e., $R_1(a,c)\wedge R_2(c,b)\Rightarrow R(a,b)$ , where $R(a,b)$ is the target and $R_2(c,b)$ is the removed fact. Table 6 shows extracted YAGO3-10 rules that are common to both models, and ones that are not. The rules show several interesting inferences, such that hasChild is often inferred via married parents, and isLocatedIn via transitivity. There are several differences in how the models reason as well; DistMult often uses the hasCapital as an intermediate step for isLocatedIn, while ConvE incorrectly uses isNeighbor. We also compare against rules extracted by BIBREF2 for YAGO3-10 that utilizes the structure of DistMult: they require domain knowledge on types and cannot be applied to ConvE. Interestingly, the extracted rules contain all the rules provided by , demonstrating that can be used to accurately interpret models, including ones that are not interpretable, such as ConvE. These are preliminary steps toward interpretability of link prediction models, and we leave more analysis of interpretability to future work. ### Finding Errors in Knowledge Graphs
Here, we demonstrate another potential use of adversarial modifications: finding erroneous triples in the knowledge graph. Intuitively, if there is an error in the graph, the triple is likely to be inconsistent with its neighborhood, and thus the model should put least trust on this triple. In other words, the error triple should have the least influence on the model's prediction of the training data. Formally, to find the incorrect triple $\langle s^{\prime }, r^{\prime }, o\rangle $ in the neighborhood of the train triple $\langle s, r, o\rangle $ , we need to find the triple $\langle s^{\prime },r^{\prime },o\rangle $ that results in the least change $\Delta _{(s^{\prime },r^{\prime })}(s,r,o)$ when removed from the graph. To evaluate this application, we inject random triples into the graph, and measure the ability of to detect the errors using our optimization. We consider two types of incorrect triples: 1) incorrect triples in the form of $\langle s^{\prime }, r, o\rangle $ where $s^{\prime }$ is chosen randomly from all of the entities, and 2) incorrect triples in the form of $\langle s^{\prime }, r^{\prime }, o\rangle $ where $s^{\prime }$ and $r^{\prime }$ are chosen randomly. We choose 100 random triples from the observed graph, and for each of them, add an incorrect triple (in each of the two scenarios) to its neighborhood. Then, after retraining DistMult on this noisy training data, we identify error triples through a search over the neighbors of the 100 facts. The result of choosing the neighbor with the least influence on the target is provided in the Table 7 . When compared with baselines that randomly choose one of the neighbors, or assume that the fact with the lowest score is incorrect, we see that outperforms both of these with a considerable gap, obtaining an accuracy of $42\%$ and $55\%$ in detecting errors. ### Related Work
Learning relational knowledge representations has been a focus of active research in the past few years, but to the best of our knowledge, this is the first work on conducting adversarial modifications on the link prediction task. Knowledge graph embedding There is a rich literature on representing knowledge graphs in vector spaces that differ in their scoring functions BIBREF21 , BIBREF22 , BIBREF23 . Although is primarily applicable to multiplicative scoring functions BIBREF0 , BIBREF1 , BIBREF2 , BIBREF24 , these ideas apply to additive scoring functions BIBREF18 , BIBREF6 , BIBREF7 , BIBREF25 as well, as we show in Appendix "First-order Approximation of the Change For TransE" . Furthermore, there is a growing body of literature that incorporates an extra types of evidence for more informed embeddings such as numerical values BIBREF26 , images BIBREF27 , text BIBREF28 , BIBREF29 , BIBREF30 , and their combinations BIBREF31 . Using , we can gain a deeper understanding of these methods, especially those that build their embeddings wit hmultiplicative scoring functions. Interpretability and Adversarial Modification There has been a significant recent interest in conducting an adversarial attacks on different machine learning models BIBREF16 , BIBREF32 , BIBREF33 , BIBREF34 , BIBREF35 , BIBREF36 to attain the interpretability, and further, evaluate the robustness of those models. BIBREF20 uses influence function to provide an approach to understanding black-box models by studying the changes in the loss occurring as a result of changes in the training data. In addition to incorporating their established method on KGs, we derive a novel approach that differs from their procedure in two ways: (1) instead of changes in the loss, we consider the changes in the scoring function, which is more appropriate for KG representations, and (2) in addition to searching for an attack, we introduce a gradient-based method that is much faster, especially for “adding an attack triple” (the size of search space make the influence function method infeasible). Previous work has also considered adversaries for KGs, but as part of training to improve their representation of the graph BIBREF37 , BIBREF38 . Adversarial Attack on KG Although this is the first work on adversarial attacks for link prediction, there are two approaches BIBREF39 , BIBREF17 that consider the task of adversarial attack on graphs. There are a few fundamental differences from our work: (1) they build their method on top of a path-based representations while we focus on embeddings, (2) they consider node classification as the target of their attacks while we attack link prediction, and (3) they conduct the attack on small graphs due to restricted scalability, while the complexity of our method does not depend on the size of the graph, but only the neighborhood, allowing us to attack real-world graphs. ### Conclusions
Motivated by the need to analyze the robustness and interpretability of link prediction models, we present a novel approach for conducting adversarial modifications to knowledge graphs. We introduce , completion robustness and interpretability via adversarial graph edits: identifying the fact to add into or remove from the KG that changes the prediction for a target fact. uses (1) an estimate of the score change for any target triple after adding or removing another fact, and (2) a gradient-based algorithm for identifying the most influential modification. We show that can effectively reduce ranking metrics on link prediction models upon applying the attack triples. Further, we incorporate the to study the interpretability of KG representations by summarizing the most influential facts for each relation. Finally, using , we introduce a novel automated error detection method for knowledge graphs. We have release the open-source implementation of our models at: https://pouyapez.github.io/criage. ### Acknowledgements
We would like to thank Matt Gardner, Marco Tulio Ribeiro, Zhengli Zhao, Robert L. Logan IV, Dheeru Dua and the anonymous reviewers for their detailed feedback and suggestions. This work is supported in part by Allen Institute for Artificial Intelligence (AI2) and in part by NSF awards #IIS-1817183 and #IIS-1756023. The views expressed are those of the authors and do not reflect the official policy or position of the funding agencies. ### Appendix
We approximate the change on the score of the target triple upon applying attacks other than the $\langle s^{\prime }, r^{\prime }, o \rangle $ ones. Since each relation appears many times in the training triples, we can assume that applying a single attack will not considerably affect the relations embeddings. As a result, we just need to study the attacks in the form of $\langle s, r^{\prime }, o \rangle $ and $\langle s, r^{\prime }, o^{\prime } \rangle $ . Defining the scoring function as $\psi (s,r,o) = , ) \cdot = _{s,r} \cdot $ , we further assume that $\psi (s,r,o) =\cdot (, ) =\cdot _{r,o}$ . ### Modifications of the Form 〈s,r ' ,o ' 〉\langle s, r^{\prime }, o^{\prime } \rangle
Using similar argument as the attacks in the form of $\langle s^{\prime }, r^{\prime }, o \rangle $ , we can calculate the effect of the attack, $\overline{\psi }{(s,r,o)}-\psi (s, r, o)$ as: (s,r,o)-(s, r, o)=(-) s, r where $_{s, r} = (,)$ . We now derive an efficient computation for $(-)$ . First, the derivative of the loss $(\overline{G})= (G)+(\langle s, r^{\prime }, o^{\prime } \rangle )$ over $$ is: es (G) = es (G) - (1-) r', o' where $_{r^{\prime }, o^{\prime }} = (^{\prime },^{\prime })$ , and $\varphi = \sigma (\psi (s,r^{\prime },o^{\prime }))$ . At convergence, after retraining, we expect $\nabla _{e_s} (\overline{G})=0$ . We perform first order Taylor approximation of $\nabla _{e_s} (\overline{G})$ to get: 0 - (1-)r',o'+ (Hs+(1-)r',o' r',o')(-) where $H_s$ is the $d\times d$ Hessian matrix for $s$ , i.e. second order derivative of the loss w.r.t. $$ , computed sparsely. Solving for $-$ gives us: -= (1-) (Hs + (1-) r',o'r',o')-1 r',o' In practice, $H_s$ is positive definite, making $H_s + \varphi (1-\varphi ) _{r^{\prime },o^{\prime }}^\intercal _{r^{\prime },o^{\prime }}$ positive definite as well, and invertible. Then, we compute the score change as: (s,r,o)-(s, r, o)= r,o (-) = ((1-) (Hs + (1-) r',o'r',o')-1 r',o')r,o. ### Modifications of the Form 〈s,r ' ,o〉\langle s, r^{\prime }, o \rangle
In this section we approximate the effect of attack in the form of $\langle s, r^{\prime }, o \rangle $ . In contrast to $\langle s^{\prime }, r^{\prime }, o \rangle $ attacks, for this scenario we need to consider the change in the $$ , upon applying the attack, in approximation of the change in the score as well. Using previous results, we can approximate the $-$ as: -= (1-) (Ho + (1-) s,r's,r')-1 s,r' and similarly, we can approximate $-$ as: -= (1-) (Hs + (1-) r',or',o)-1 r',o where $H_s$ is the Hessian matrix over $$ . Then using these approximations: s,r(-) = s,r ((1-) (Ho + (1-) s,r's,r')-1 s,r') and: (-) r,o= ((1-) (Hs + (1-) r',or',o)-1 r',o) r,o and then calculate the change in the score as: (s,r,o)-(s, r, o)= s,r.(-) +(-).r,o = s,r ((1-) (Ho + (1-) s,r's,r')-1 s,r')+ ((1-) (Hs + (1-) r',or',o)-1 r',o) r, o ### First-order Approximation of the Change For TransE
In here we derive the approximation of the change in the score upon applying an adversarial modification for TransE BIBREF18 . Using similar assumptions and parameters as before, to calculate the effect of the attack, $\overline{\psi }{(s,r,o)}$ (where $\psi {(s,r,o)}=|+-|$ ), we need to compute $$ . To do so, we need to derive an efficient computation for $$ . First, the derivative of the loss $(\overline{G})= (G)+(\langle s^{\prime }, r^{\prime }, o \rangle )$ over $$ is: eo (G) = eo (G) + (1-) s', r'-(s',r',o) where $_{s^{\prime }, r^{\prime }} = ^{\prime }+ ^{\prime }$ , and $\varphi = \sigma (\psi (s^{\prime },r^{\prime },o))$ . At convergence, after retraining, we expect $\nabla _{e_o} (\overline{G})=0$ . We perform first order Taylor approximation of $\nabla _{e_o} (\overline{G})$ to get: 0 (1-) (s', r'-)(s',r',o)+(Ho - Hs',r',o)(-) Hs',r',o = (1-)(s', r'-)(s', r'-)(s',r',o)2+ 1-(s',r',o)-(1-) (s', r'-)(s', r'-)(s',r',o)3 where $H_o$ is the $d\times d$ Hessian matrix for $o$ , i.e., second order derivative of the loss w.r.t. $$ , computed sparsely. Solving for $$ gives us: = -(1-) (Ho - Hs',r',o)-1 (s', r'-)(s',r',o) + Then, we compute the score change as: (s,r,o)= |+-| = |++(1-) (Ho - Hs',r',o)-1 (s', r'-)(s',r',o) - | Calculating this expression is efficient since $H_o$ is a $d\times d$ matrix. ### Sample Adversarial Attacks
In this section, we provide the output of the for some target triples. Sample adversarial attacks are provided in Table 5 . As it shows, attacks mostly try to change the type of the target triple's object by associating it with a subject and a relation that require a different entity types. Figure 1: Completion Robustness and Interpretability via Adversarial Graph Edits (CRIAGE): Change in the graph structure that changes the prediction of the retrained model, where (a) is the original sub-graph of the KG, (b) removes a neighboring link of the target, resulting in a change in the prediction, and (c) shows the effect of adding an attack triple on the target. These modifications were identified by our proposed approach. Figure 2: Inverter Network The architecture of our inverter function that translate zs,r to its respective (s̃, r̃). The encoder component is fixed to be the encoder network of DistMult and ConvE respectively. Table 1: Inverter Functions Accuracy, we calculate the accuracy of our inverter networks in correctly recovering the pairs of subject and relation from the test set of our benchmarks. Table 2: Data Statistics of the benchmarks. Figure 3: Influence function vs CRIAGE. We plot the average time (over 10 facts) of influence function (IF) and CRIAGE to identify an adversary as the number of entities in the Kinship KG is varied (by randomly sampling subgraphs of the KG). Even with small graphs and dimensionality, IF quickly becomes impractical. Table 3: Ranking modifications by their impact on the target. We compare the true ranking of candidate triples with a number of approximations using ranking correlation coefficients. We compare our method with influence function (IF) with and without Hessian, and ranking the candidates based on their score, on two KGs (d = 10, averaged over 10 random targets). For the sake of brevity, we represent the Spearman’s ρ and Kendall’s τ rank correlation coefficients simply as ρ and τ . Table 4: Robustness of Representation Models, the effect of adversarial attack on link prediction task. We consider two scenario for the target triples, 1) choosing the whole test dataset as the targets (All-Test) and 2) choosing a subset of test data that models are uncertain about them (Uncertain-Test). Figure 4: Per-Relation Breakdown showing the effect of CRIAGE-Add on different relations in YAGO3-10. Table 5: Extracted Rules for identifying the most influential link. We extract the patterns that appear more than 90% times in the neighborhood of the target triple. The output of CRIAGE-Remove is presented in red. Table 6: Error Detection Accuracy in the neighborhood of 100 chosen samples. We choose the neighbor with the least value of ∆(s′,r′)(s, r, o) as the incorrect fact. This experiment assumes we know each target fact has exactly one error. Table 7: Top adversarial triples for target samples. | if there is an error in the graph, the triple is likely to be inconsistent with its neighborhood, and thus the model should put least trust on this triple. In other words, the error triple should have the least influence on the model's prediction of the training data. |
What was Orison’s excuse to visit the upper floors?
A. To deliver a message from Mr. Wanji
B. To see what the Earmuffs were doing
C. To feed the Microfabridae
D. To complain about Auga Vingt
| CINDERELLA STORY By ALLEN KIM LANG What a bank! The First Vice-President was a cool cat—the elevator and the money operators all wore earmuffs—was just as phony as a three-dollar bill! [Transcriber's Note: This etext was produced from Worlds of If Science Fiction, May 1961. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed.] I The First Vice-President of the William Howard Taft National Bank and Trust Company, the gentleman to whom Miss Orison McCall was applying for a job, was not at all the public picture of a banker. His suit of hound's-tooth checks, the scarlet vest peeping above the vee of his jacket, were enough to assure Orison that the Taft Bank was a curious bank indeed. "I gotta say, chick, these references of yours really swing," said the Vice-President, Mr. Wanji. "Your last boss says you come on real cool in the secretary-bit." "He was a very kind employer," Orison said. She tried to keep from staring at the most remarkable item of Mr. Wanji's costume, a pair of furry green earmuffs. It was not cold. Mr. Wanji returned to Orison her letters of reference. "What color bread you got eyes for taking down, baby?" he asked. "Beg pardon?" "What kinda salary you bucking for?" he translated, bouncing up and down on the toes of his rough-leather desert boots. "I was making one-twenty a week in my last position," Miss McCall said. "You're worth more'n that, just to jazz up the decor," Mr. Wanji said. "What you say we pass you a cee-and-a-half a week. Okay?" He caught Orison's look of bewilderment. "One each, a Franklin and a Grant," he explained further. She still looked blank. "Sister, you gonna work in a bank, you gotta know who's picture's on the paper. That's a hunnerd-fifty a week, doll." "That will be most satisfactory, Mr. Wanji," Orison said. It was indeed. "Crazy!" Mr. Wanji grabbed Orison's right hand and shook it with athletic vigor. "You just now joined up with our herd. I wanna tell you, chick, it's none too soon we got some decent scenery around this tomb, girlwise." He took her arm and led her toward the bank of elevators. The uniformed operator nodded to Mr. Wanji, bowed slightly to Orison. He, too, she observed, wore earmuffs. His were more formal than Mr. Wanji's, being midnight blue in color. "Lift us to five, Mac," Mr. Wanji said. As the elevator door shut he explained to Orison, "You can make the Taft Bank scene anywhere between the street floor and floor five. Basement and everything higher'n fifth floor is Iron Curtain Country far's you're concerned. Dig, baby?" "Yes, sir," Orison said. She was wondering if she'd be issued earmuffs, now that she'd become an employee of this most peculiar bank. The elevator opened on five to a tiny office, just large enough to hold a single desk and two chairs. On the desk were a telephone and a microphone. Beside them was a double-decked "In" and "Out" basket. "Here's where you'll do your nine-to-five, honey," Mr. Wanji said. "What will I be doing, Mr. Wanji?" Orison asked. The Vice-President pointed to the newspaper folded in the "In" basket. "Flip on the microphone and read the paper to it," he said. "When you get done reading the paper, someone will run you up something new to read. Okay?" "It seems a rather peculiar job," Orison said. "After all, I'm a secretary. Is reading the newspaper aloud supposed to familiarize me with the Bank's operation?" "Don't bug me, kid," Mr. Wanji said. "All you gotta do is read that there paper into this here microphone. Can do?" "Yes, sir," Orison said. "While you're here, Mr. Wanji, I'd like to ask you about my withholding tax, social security, credit union, coffee-breaks, union membership, lunch hour and the like. Shall we take care of these details now? Or would you—" "You just take care of that chicken-flickin' kinda stuff any way seems best to you, kid," Mr. Wanji said. "Yes, sir," Orison said. This laissez-faire policy of Taft Bank's might explain why she'd been selected from the Treasury Department's secretarial pool to apply for work here, she thought. Orison McCall, girl Government spy. She picked up the newspaper from the "In" basket, unfolded it to discover the day's Wall Street Journal , and began at the top of column one to read it aloud. Wanji stood before the desk, nodding his head as he listened. "You blowing real good, kid," he said. "The boss is gonna dig you the most." Orison nodded. Holding her newspaper and her microphone, she read the one into the other. Mr. Wanji flicked his fingers in a good-by, then took off upstairs in the elevator. By lunchtime Orison had finished the Wall Street Journal and had begun reading a book an earmuffed page had brought her. The book was a fantastic novel of some sort, named The Hobbit . Reading this peculiar fare into the microphone before her, Miss McCall was more certain than ever that the Taft Bank was, as her boss in Washington had told her, the front for some highly irregular goings-on. An odd business for a Federal Mata Hari, Orison thought, reading a nonsense story into a microphone for an invisible audience. Orison switched off her microphone at noon, marked her place in the book and took the elevator down to the ground floor. The operator was a new man, ears concealed behind scarlet earmuffs. In the car, coming down from the interdicted upper floors, were several gentlemen with briefcases. As though they were members of a ballet-troupe, these gentlemen whipped off their hats with a single motion as Orison stepped aboard the elevator. Each of the chivalrous men, hat pressed to his heart, wore a pair of earmuffs. Orison nodded bemused acknowledgment of their gesture, and got off in the lobby vowing never to put a penny into this curiousest of banks. Lunch at the stand-up counter down the street was a normal interlude. Girls from the ground-floor offices of Taft Bank chattered together, eyed Orison with the coolness due so attractive a competitor, and favored her with no gambit to enter their conversations. Orison sighed, finished her tuna salad on whole-wheat, then went back upstairs to her lonely desk and her microphone. By five, Orison had finished the book, reading rapidly and becoming despite herself engrossed in the saga of Bilbo Baggins, Hobbit. She switched off the microphone, put on her light coat, and rode downstairs in an elevator filled with earmuffed, silent, hat-clasping gentlemen. What I need, Orison thought, walking rapidly to the busline, is a double Scotch, followed by a double Scotch. And what the William Howard Taft National Bank and Trust Company needs is a joint raid by forces of the U.S. Treasury Department and the American Psychiatric Association. Earmuffs, indeed. Fairy-tales read into a microphone. A Vice-President with the vocabulary of a racetrack tout. And what goes on in those upper floors? Orison stopped in at the restaurant nearest her apartment house—the Windsor Arms—and ordered a meal and a single Martini. Her boss in Washington had told her that this job of hers, spying on Taft Bank from within, might prove dangerous. Indeed it was, she thought. She was in danger of becoming a solitary drinker. Home in her apartment, Orison set the notes of her first day's observations in order. Presumably Washington would call tonight for her initial report. Item: some of the men at the Bank wore earmuffs, several didn't. Item: the Vice-President's name was Mr. Wanji: Oriental? Item: the top eight floors of the Taft Bank Building seemed to be off-limits to all personnel not wearing earmuffs. Item: she was being employed at a very respectable salary to read newsprint and nonsense into a microphone. Let Washington make sense of that, she thought. In a gloomy mood, Orison McCall showered and dressed for bed. Eleven o'clock. Washington should be calling soon, inquiring after the results of her first day's spying. No call. Orison slipped between the sheets at eleven-thirty. The clock was set; the lights were out. Wasn't Washington going to call her? Perhaps, she thought, the Department had discovered that the Earmuffs had her phone tapped. "Testing," a baritone voice muttered. Orison sat up, clutching the sheet around her throat. "Beg pardon?" she said. "Testing," the male voice repeated. "One, two, three; three, two, one. Do you read me? Over." Orison reached under the bed for a shoe. Gripping it like a Scout-ax, she reached for the light cord with her free hand and tugged at it. The room was empty. "Testing," the voice repeated. "What you're testing," Orison said in a firm voice, "is my patience. Who are you?" "Department of Treasury Monitor J-12," the male voice said. "Do you have anything to report, Miss McCall?" "Where are you, Monitor?" she demanded. "That's classified information," the voice said. "Please speak directly to your pillow, Miss McCall." Orison lay down cautiously. "All right," she whispered to her pillow. "Over here," the voice instructed her, coming from the unruffled pillow beside her. Orison transferred her head to the pillow to her left. "A radio?" she asked. "Of a sort," Monitor J-12 agreed. "We have to maintain communications security. Have you anything to report?" "I got the job," Orison said. "Are you ... in that pillow ... all the time?" "No, Miss McCall," the voice said. "Only at report times. Shall we establish our rendezvous here at eleven-fifteen, Central Standard Time, every day?" "You make it sound so improper," Orison said. "I'm far enough away to do you no harm, Miss McCall," the monitor said. "Now, tell me what happened at the bank today." Orison briefed her pillow on the Earmuffs, on her task of reading to a microphone, and on the generally mimsy tone of the William Howard Taft National Bank and Trust Company. "That's about it, so far," she said. "Good report," J-12 said from the pillow. "Sounds like you've dropped into a real snakepit, beautiful." "How do you know ... why do you think I'm beautiful?" Orison asked. "Native optimism," the voice said. "Good night." J-12 signed off with a peculiar electronic pop that puzzled Orison for a moment. Then she placed the sound: J-12 had kissed his microphone. Orison flung the shoe and the pillow under her bed, and resolved to write Washington for permission to make her future reports by registered mail. II At ten o'clock the next morning, reading page four of the current Wall Street Journal , Orison was interrupted by the click of a pair of leather heels. The gentleman whose heels had just slammed together was bowing. And she saw with some gratification that he was not wearing earmuffs. "My name," the stranger said, "is Dink Gerding. I am President of this bank, and wish at this time to welcome you to our little family." "I'm Orison McCall," she said. A handsome man, she mused. Twenty-eight? So tall. Could he ever be interested in a girl just five-foot-three? Maybe higher heels? "We're pleased with your work, Miss McCall," Dink Gerding said. He took the chair to the right of her desk. "It's nothing," Orison said, switching off the microphone. "On the contrary, Miss McCall. Your duties are most important," he said. "Reading papers and fairy-tales into this microphone is nothing any reasonably astute sixth-grader couldn't do as well," Orison said. "You'll be reading silently before long," Mr. Gerding said. He smiled, as though this explained everything. "By the way, your official designation is Confidential Secretary. It's me whose confidences you're to keep secret. If I ever need a letter written, may I stop down here and dictate it?" "Please do," Orison said. This bank president, for all his grace and presence, was obviously as kookie as his bank. "Have you ever worked in a bank before, Miss McCall?" Mr. Gerding asked, as though following her train of thought. "No, sir," she said. "Though I've been associated with a rather large financial organization." "You may find some of our methods a little strange, but you'll get used to them," he said. "Meanwhile, I'd be most grateful if you'd dispense with calling me 'sir.' My name is Dink. It is ridiculous, but I'd enjoy your using it." "Dink?" she asked. "And I suppose you're to call me Orison?" "That's the drill," he said. "One more question, Orison. Dinner this evening?" Direct, she thought. Perhaps that's why he's president of a bank, and still so young. "We've hardly met," she said. "But we're on a first-name basis already," he pointed out. "Dance?" "I'd love to," Orison said, half expecting an orchestra to march, playing, from the elevator. "Then I'll pick you up at seven. Windsor Arms, if I remember your personnel form correctly." He stood, lean, all bone and muscle, and bowed slightly. West Point? Hardly. His manners were European. Sandhurst, perhaps, or Saint Cyr. Was she supposed to reply with a curtsy? Orison wondered. "Thank you," she said. He was a soldier, or had been: the way, when he turned, his shoulders stayed square. The crisp clicking of his steps, a military metronome, to the elevator. When the door slicked open Orison, staring after Dink, saw that each of the half-dozen men aboard snapped off their hats (but not their earmuffs) and bowed, the earmuffed operator bowing with them. Small bows, true; just head-and-neck. But not to her. To Dink Gerding. Orison finished the Wall Street Journal by early afternoon. A page came up a moment later with fresh reading-matter: a copy of yesterday's Congressional Record . She launched into the Record , thinking as she read of meeting again this evening that handsome madman, that splendid lunatic, that unlikely bank-president. "You read so well , darling," someone said across the desk. Orison looked up. "Oh, hello," she said. "I didn't hear you come up." "I walk ever so lightly," the woman said, standing hip-shot in front of the desk, "and pounce ever so hard." She smiled. Opulent, Orison thought. Built like a burlesque queen. No, she thought, I don't like her. Can't. Wouldn't if I could. Never cared for cats. "I'm Orison McCall," she said, and tried to smile back without showing teeth. "Delighted," the visitor said, handing over an undelighted palm. "I'm Auga Vingt. Auga, to my friends." "Won't you sit down, Miss Vingt?" "So kind of you, darling," Auga Vingt said, "but I shan't have time to visit. I just wanted to stop and welcome you as a Taft Bank co-worker. One for all, all for one. Yea, Team. You know." "Thanks," Orison said. "Common courtesy," Miss Vingt explained. "Also, darling, I'd like to draw your attention to one little point. Dink Gerding—you know, the shoulders and muscles and crewcut? Well, he's posted property. Should you throw your starveling charms at my Dink, you'd only get your little eyes scratched out. Word to the wise, n'est-ce pas ?" "Sorry you have to leave so suddenly," Orison said, rolling her Wall Street Journal into a club and standing. "Darling." "So remember, Tiny, Dink Gerding is mine. You're all alone up here. You could get broken nails, fall down the elevator shaft, all sorts of annoyance. Understand me, darling?" "You make it very clear," Orison said. "Now you'd best hurry back to your stanchion, Bossy, before the hay's all gone." "Isn't it lovely, the way you and I reached an understanding right off?" Auga asked. "Well, ta-ta." She turned and walked to the elevator, displaying, Orison thought, a disgraceful amount of ungirdled rhumba motion. The elevator stopped to pick up the odious Auga. A passenger, male, stepped off. "Good morning, Mr. Gerding," Miss Vingt said, bowing. "Carry on, Colonel," the stranger replied. As the elevator door closed, he stepped up to Orison's desk. "Good morning. Miss McCall," he said. "What is this?" Orison demanded. "Visiting-day at the zoo?" She paused and shook her head. "Excuse me, sir," she said. "It's just that ... Vingt thing...." "Auga is rather intense," the new Mr. Gerding said. "Yeah, intense," Orison said. "Like a kidney-stone." "I stopped by to welcome you to the William Howard Taft National Bank and Trust Company family, Miss McCall," he said. "I'm Kraft Gerding, Dink's elder brother. I understand you've met Dink already." "Yes, sir," Orison said. The hair of this new Mr. Gerding was cropped even closer than Dink's. His mustache was gray-tipped, like a patch of frosted furze; and his eyes, like Dink's, were cobalt blue. The head, Orison mused, would look quite at home in one of Kaiser Bill's spike-topped Pickelhauben ; but the ears were in evidence, and seemed normal. Mr. Kraft Gerding bowed—what continental manners these bankers had!—and Orison half expected him to free her hand from the rolled-up paper she still clutched and plant a kiss on it. Instead, Kraft Gerding smiled a smile as frosty as his mustache and said, "I understand that my younger brother has been talking with you, Miss McCall. Quite proper, I know. But I must warn you against mixing business with pleasure." Orison jumped up, tossing the paper into her wastebasket. "I quit!" she shouted. "You can take this crazy bank ... into bankruptcy, for all I care. I'm not going to perch up here, target for every uncaged idiot in finance, and listen to another word." "Dearest lady, my humblest pardon," Kraft Gerding said, bowing again, a bit lower. "Your work is splendid; your presence is Taft Bank's most charming asset; my only wish is to serve and protect you. To this end, dear lady, I feel it my duty to warn you against my brother. A word to the wise...." " N'est-ce pas? " Orison said. "Well, Buster, here's a word to the foolish. Get lost." Kraft Gerding bowed and flashed his gelid smile. "Until we meet again?" "I'll hold my breath," Orison promised. "The elevator is just behind you. Push a button, will you? And bon voyage ." Kraft Gerding called the elevator, marched aboard, favored Orison with a cold, quick bow, then disappeared into the mysterious heights above fifth floor. First the unspeakable Auga Vingt, then the obnoxious Kraft Gerding. Surely, Orison thought, recovering the Wall Street Journal from her wastebasket and smoothing it, no one would convert a major Midwestern bank into a lunatic asylum. How else, though, could the behavior of the Earmuffs be explained? Could madmen run a bank? Why not, she thought. History is rich in examples of madmen running nations, banks and all. She began again to read the paper into the microphone. If she finished early enough, she might get a chance to prowl those Off-Limits upper floors. Half an hour further into the paper, Orison jumped, startled by the sudden buzz of her telephone. She picked it up. " Wanji e-Kal, Datto. Dink ger-Dink d'summa. " Orison scribbled down this intelligence in bemused Gregg before replying, "I'm a local girl. Try me in English." "Oh. Hi, Miss McCall," the voice said. "Guess I goofed. I'm in kinda clutch. This is Wanji. I got a kite for Mr. Dink Gerding. If you see him, tell him the escudo green is pale. Got that, doll?" "Yes, Mr. Wanji. I'll tell Mr. Gerding." Orison clicked the phone down. What now, Mata Hari? she asked herself. What was the curious language Mr. Wanji had used? She'd have to report the message to Washington by tonight's pillow, and let the polyglots of Treasury Intelligence puzzle it out. Meanwhile, she thought, scooting her chair back from her desk, she had a vague excuse to prowl the upper floors. The Earmuffs could only fire her. Orison folded the paper and put it in the "Out" basket. Someone would be here in a moment with something new to read. She'd best get going. The elevator? No. The operators had surely been instructed to keep her off the upstairs floors. But the building had a stairway. III The door on the sixth floor was locked. Orison went on up the stairs to seven. The glass of the door there was painted black on the inside, and the landing was cellar-dark. Orison closed her eyes for a moment. There was a curious sound. The buzzing of a million bees, barely within the fringes of her hearing. Somehow, a very pleasant sound. She opened her eyes and tried the knob. The door opened. Orison was blinded by the lights, brilliant as noonday sun. The room extended through the entire seventh floor, its windows boarded shut, its ceiling a mass of fluorescent lamps. Set about the floor were galvanized steel tanks, rectangular and a little bigger than bathtubs. Orison counted the rows of tanks. Twelve rows, nine tiers. One hundred and eight tanks. She walked closer. The tubs were laced together by strands of angel-hair, delicate white lattices scintillating with pink. She walked to the nearest of the tubs and looked in. It was half full of a greenish fluid, seething with tiny pink bubbles. For a moment Orison thought she saw Benjamin Franklin winking up at her from the liquid. Then she screamed. The pink bubbles, the tiny flesh-colored flecks glinting light from the spun-sugar bridges between the tanks, were spiders. Millions upon millions of spiders, each the size of a mustard-seed; crawling, leaping, swinging, spinning webs, seething in the hundred tanks. Orison put her hands over her ears and screamed again, backing toward the stairway door. Into a pair of arms. "I had hoped you'd be happy here, Miss McCall," Kraft Gerding said. Orison struggled to release herself. She broke free only to have her wrists seized by two Earmuffs that had appeared with the elder Gerding. "It seems that our Pandora doesn't care for spiders," he said. "Really, Miss McCall, our little pets are quite harmless. Were we to toss you into one of these tanks...." Orison struggled against her two sumo -sized captors, whose combined weights exceeded hers by some quarter-ton, without doing more than lifting her feet from the floor. "... your flesh would be unharmed, though they spun and darted all around you. Our Microfabridae are petrovorous, Miss McCall. Of course, once they discovered your teeth, and through them a skeleton of calcium, a delicacy they find most toothsome, you'd be filleted within minutes." "Elder Compassion wouldn't like your harming the girl, Sire," one of the earmuffed sumo -wrestlers protested. "Elder Compassion has no rank," Kraft Gerding said. "Miss McCall, you must tell me what you were doing here, or I'll toss you to the spiders." "Dink ... Dink!" Orison shouted. "My beloved younger brother is otherwise engaged than in the rescue of damsels in distress," Kraft said. "Someone, after all, has to mind the bank." "I came to bring a message to Dink," Orison said. "Let me go, you acromegalic apes!" "The message?" Kraft Gerding demanded. "Something about escudo green. Put me down!" Suddenly she was dropped. Her mountainous keepers were on the floor as though struck by lightning, their arms thrown out before them, their faces abject against the floor. Kraft Gerding was slowly lowering himself to one knee. Dink had entered the spider-room. Without questions, he strode between the shiko-ing Earmuffs and put his arms around Orison. "They can't harm you," he said. She turned to press her face against his chest. "You're all right, child. Breathe deep, swallow, and turn your brain back on. All right, now?" "All right," she said, still trembling. "They were going to throw me to the spiders." "Kraft told you that?" Dink Gerding released her and turned to the kneeling man. "Stand up, Elder Brother." "I...." Dink brought his right fist up from hip-level, crashing it into Kraft's jaw. Kraft Gerding joined the Earmuffs on the floor. "If you'd care to stand again, Elder Brother, you may attempt to recover your dignity without regard for the difference in our rank." Kraft struggled to one knee and remained kneeling, gazing up at Dink through half-closed eyes. "No? Then get out of here, all of you. Samma! " Kraft Gerding arose, stared for a moment at Dink and Orison, then, with the merest hint of a bow, led his two giant Earmuffs to the elevator. "I wish you hadn't come up here, Orison," Dink said. "Why did you do it?" "Have you read the story of Bluebeard?" Orison asked. She stood close to Dink, keeping her eyes on the nearest spidertank. "I had to see what it was you kept up here so secretly, what it was that I was forbidden to see. My excuse was to have been that I was looking for you, to deliver a message from Mr. Wanji. He said I was to tell you that the escudo green is pale." "You're too curious, and Wanji is too careless," Dink said. "Now, what is this thing you have about spiders?" "I've always been terrified of them," Orison said. "When I was a little girl, I had to stay upstairs all day one Sunday because there was a spider hanging from his thread in the stairway. I waited until Dad came home and took it down with a broom. Even then, I didn't have appetite for supper." "Strange," Dink said. He walked over to the nearest tank and plucked one of the tiny pink creatures from a web-bridge. "This is no spider, Orison," he said. She backed away from Dink Gerding and the minuscule creature he cupped in the palm of his hand. "These are Microfabridae, more nearly related to shellfish than to spiders," he said. "They're stone-and-metal eaters. They literally couldn't harm a fly. Look at it, Orison." He extended his palm. Orison forced herself to look. The little creature, flesh-colored against his flesh, was nearly invisible, scuttling around the bowl of his hand. "Pretty little fellow, isn't he?" Dink asked. "Here. You hold him." "I'd rather not," she protested. "I'd be happier if you did," Dink said. Orison extended her hand as into a furnace. Dink brushed the Microfabridus from his palm to hers. It felt crisp and hard, like a legged grain of sand. Dink took a magnifier from his pocket and unfolded it, to hold it over Orison's palm. "He's like a baby crawdad," Orison said. "A sort of crustacean," Dink agreed. "We use them in a commercial process we're developing. That's why we keep this floor closed off and secret. We don't have a patent on the use of Microfabridae, you see." "What do they do?" Orison asked. "That's still a secret," Dink said, smiling. "I can't tell even you that, not yet, even though you're my most confidential secretary." "What's he doing now?" Orison asked, watching the Microfabridus, perched up on the rear four of his six microscopic legs, scratching against her high-school class-ring with his tiny chelae. "They like gold," Dink explained, peering across her shoulder, comfortably close. "They're attracted to it by a chemical tropism, as children are attracted to candy. Toss him back into his tank, Orison. We'd better get you down where you belong." Orison brushed the midget crustacean off her finger into the nearest tank, where he joined the busy boil of his fellows. She felt her ring. It was pitted where the Microfabridus had been nibbling. "Strange, using crawdads in a bank," she said. She stood silent for a moment. "I thought I heard music," she said. "I heard it when I came in. Something like the sighing of wind in winter trees." "That's the hymn of the Microfabridae," Dink said. "They all sing together while they work, a chorus of some twenty million voices." He took her arm. "If you listen very carefully, you'll find the song these little workers sing the most beautiful music in the world." Orison closed her eyes, leaning back into Dink's arms, listening to the music that seemed on the outermost edge of her hearing. Wildness, storm and danger were its theme, counterpointed by promises of peace and harbor. She heard the wash of giant waves in the song, the crash of breakers against granite, cold and insatiable. And behind this, the quiet of sheltered tide-pools, the soft lub of sea-arms landlocked. "It's an ancient song," Dink said. "The Microfabridae have been singing it for a million years." He released her, and opened a wood-covered wooden box. He scooped up a cupful of the sand inside. "Hold out your hands," he told Orison. He filled them with the sand. "Throw our singers some supper for their song," he said. Orison went with her cupped hands to the nearest tank and sprinkled the mineral fishfood around inside it. The Microfabridae leaped from the liquid like miniature porpoises, seizing the grains of sand in mid-air. "They're so very strange," Orison said. At the bottom of the tank she thought she saw Ben Franklin again, winking at her through the bubbling life. Nonsense, she thought, brushing her hands. | A. To deliver a message from Mr. Wanji |
What isn't a way that the Times treated Dole unfairly?
A. they published unfavorable pictures of him
B. the way they quoted him emphasized his flaws
C. they had no full-time reporters following him
D. they omitted information about Dole's successes
| Dole vs. the Times For several weeks now, pundits have debated how Bob Dole would exit the stage. Would he depart on a negative note about his opponent or a positive one about himself? Would he leave with anger or with humor? In the past several days, the issue has been settled. Dole, it appears, will end his political career raging against the New York Times . Dole's spat with the gray lady went public on Thursday, Oct. 24. In New Orleans, Dole charged the paper with ignoring a story about a Miami drug dealer who got invited to the White House. "This is a disgrace," Dole insisted. "I doubt if you even read it in the New York Times . They probably put it in the want ads. They don't put any anti-Clinton stories in the New York Times . Only anti-Dole stories in the New York Times ." Dole repeated his attack for the next five days. "We are not going to let the media steal this election," he told a crowd in Dallas on Friday. "This country belongs to the people, not the New York Times ." On Saturday, in Visalia, Calif., he added, "I know that with a crowd this size, the New York Times will write not many people showed up, but the other papers will get it right." On Sunday (the day the Times endorsed Clinton), Dole called the paper "the apologist for President Clinton for the last four years and an arm of the Democratic National Committee." In a CNN interview broadcast Monday, Dole said the Times "might as well be part of the Democratic Party. ... They hammer us on a daily basis. We make a major speech, they bury it back on section D. They put a front-page story that, well, Bob Dole and Jack Kemp didn't get along together 12 years ago." On Tuesday, Dole was still at it, referring to the 28 words of the 10th Amendment, and quipping, "That's about what I got in the New York Times today." The Times has reacted to this assault by highhandedly quoting everything and explaining none of it, leaving its readers baffled as to why the Republican nominee is so upset at the paper. In fact, Dole's fury at the Times is hardly news to those who work at the paper. According to Katharine Seelye, who has covered Dole since the beginning of his campaign, the complaints date from December 1995, when Dole staff members first protested that she had misunderstood the candidate's position on abortion. The real bitterness, however, began in May, when the paper played what Dole aides billed as a major address about welfare on Page 19 of the business section. Since then, campaign honchos have peppered the paper's reporters and editors with constant phone calls and letters complaining about unfair treatment. Reporters traveling with Dole caught a glimpse of the enmity Oct. 9, when Nelson Warfield, Dole's press secretary, staged a public confrontation with Seelye. The candidate, Warfield told reporters waiting to board the campaign plane, had just come from an appearance on G. Gordon Liddy's radio show. Why, Seelye asked, weren't reporters told about the appearance in advance? According to reporters present, Warfield snapped that it wouldn't make any difference because the Times would get the story wrong anyway. Then, on the plane, Warfield walked back to the press section and grandly served Seelye with a copy of a letter from Communications Director John Buckley to her boss, Times Washington Editor Andrew Rosenthal. That letter, which has fallen into the hands of Slate, protests Seelye's coverage of a speech the previous day. Dole, in New Jersey, had talked about Clinton being AWOL in the drug war. "Where has he been for four years? How many hundreds of thousands of young people started drugs?" Dole said. "Three million have started smoking while he was playing around with smoking and all this stuff finally in an election year." Seelye's front-page story reported that "Mr. Dole accused the President of 'playing around' while the drug war raged out of control." Buckley complains that the story "could lead the reader to believe that Dole was talking about a very different kind of 'playing around'--something he did not say, and something he would not say." The letter continues: "Since May, I have been pointing out to you a problem we see with the accuracy and understanding of context revealed in Kit's reporting," going on to assert that "Seelye has misquoted Dole on numerous occasions and done so in a manner that distorted the accuracy of her assertions and your coverage." No Dole staff would be quoted by name for this story, but speaking on background, a senior campaign official elaborated upon the complaint. "They've just done a miserable job throughout this campaign," the official said. "The coverage of Dole has been excessively bitchy from day one, in addition to having a number of extraordinary factual problems." With Seelye, the official says, the problem is "not being able to transcribe a tape accurately." With Adam Nagourney, the Times ' other reporter covering Dole full time since the summer, "the problem is an incredible focus on the little picture as opposed to the big picture." As an example, the official cites a September story in which Nagourney lumped together Dole's fall from a platform in Chico, Calif., and his mistaken reference to the "Brooklyn" Dodgers as "a rough stretch of politicking." Other than those two episodes, the official says, Dole actually had a great week. The campaign's complaint extends to unequal treatment--a nine-part series on Clinton's record, which the official describes as "the softest portrait since they invented black velvet"--and the Times perpetually underestimating the size of Dole crowds. "Clinton even gets better photographs," the official contends. Rosenthal, who has direct responsibility for campaign coverage at the Times , professes bewilderment at these complaints. "We don't make editorial judgments based on disposition to be tough on Bob Dole or nice to Bob Dole," he says. On the specifics, Rosenthal says that the Times ran an editor's note acknowledging that it shouldn't have truncated the "playing around" quote. He points out that the Times ran its story on the Miami drug dealer who visited the White House the same day Dole accused the paper of not covering it. As for the nine-part series on Clinton, Rosenthal says it is the long-standing practice of the paper to do a lengthy series on the incumbent's record. "If Dole wins and runs again in 2000, he will get nine-part series too," he says. "Ithink we have been tough on him," Seelye says. This stems, however, not from any bias, she says, but from the campaign's own internal problems. Dole's campaign has been especially "porous," with aides emulating the proverbial seafaring rats. This is true enough--in recent days ex-strategist Don Sipple has trashed the campaign on the record. But there's another point, too. Contrary to Buckley's charge that she misquotes Dole, Seelye routinely makes Dole look ridiculous by quoting him all too accurately, depicting him in what one colleague calls a "cinema verité " style. Famous for going over and over her tape recordings on the campaign plane, Seelye manages to get every Dole mumble, repetition, and verbal miscue down. For instance, in her Oct. 26 story reporting Dole's attack on the Times , Seelye writes: "In Phoenix on Friday night, he had a delightful time drawing out his vowels as he described financial contributions to the Clinton campaign. "From Indoneeesia," he said. "Yeah. From INdiaaaaah. Some fellow named Gandhi out there. He owes $10,000 in back taxes, but he found $300,000 to give to the Clinton campaign. And now Gandhi is gaaaawn. Gaaaaandhi, gone gone gone. They can't find him." Two days later, she quoted Dole in another story: "They've turned the White House into something else, I don't know what it is. It's the animal house! It's the animal house!" Most reporters would write, Bob Dole yesterday compared the White House to an "animal house," sparing the exclamation points, and making him sound at least compos mentis. But though unflattering, Seelye's Mametizing of Bob Dole can hardly be called unfair. It is not as if the Times cleans up Clinton's quotes; the president simply observes the rules of syntax most of the time. Something similar may be happening with the pictures. After four years, Clinton has learned how to avoid looking unpresidential. He no longer allows himself to be photographed wearing too-short running shorts, and he avoids pulling faces in public. Dole, who is simply less photogenic, is an easier victim for picture editors--who, like their editorial counterparts, have a strong bias against dullness. Take, for instance, the two pictures shown above. The front-page picture the Times ran the day after the second presidential debate does make Dole look like a decomposing monster. But unlike the picture in the Washington Post the same day, it captures the spirit of the event, with Dole grimly taking the offensive and Clinton watching warily but standing aside from the attacks. Dole sounds absurd when he alleges that the paper that broke Whitewater and the story of the first lady's commodities trades has not been aggressive in pursuing Clinton scandals. All sorts of potential Dole scandals have been soft-pedaled by the media, including the Times , because he is so far behind. It's true that coverage of Clinton on the campaign trail has been somewhat softer than the coverage of Dole, as even other Times reporters acknowledge. But the explanation is institutional, not ideological. The press, as many have complained, overemphasizes the "horse race" aspect of politics. As a side effect of that disease, reporters have excessive respect for a well-run campaign. (In 1988, Republican George Bush benefited from this phenomenon.) A cruder reality is that reporters need to have a relationship with Clinton after Tuesday. None of these factors, though, is unique to the Times . So why is Dole singling it out? Dole's attacks on the Times have the appearance of being an exercise in populist demagogy. In one of his great cue-card reading remarks, Dole tried to explain his recent attacks on CNN the other night by saying, "I like the media. They don't like them in the South." But this pat explanation doesn't entirely make sense. Red meat for right-wing crowds doesn't help Dole with the centrist voters he would need to turn around in order to make the miraculous happen. And in fact, according to a senior Dole aide, the attacks are heartfelt on the candidate's part. Dole has been going after the Times over the objections of advisers who have been telling him there's no percentage in picking fights with the press. But if Dole is attacking the Times because he is truly furious and not because he thinks it will help him get elected, what is he so angry about? The answer, I think, is that there has always been a Nixonian streak in Bob Dole, by which I mean a part of him which feels shut out of the closed circle of the Eastern establishment. At the Republican convention, Dole blasted the Clinton administration as a "corps of the elite who never grew up, never did anything real, never sacrificed, never suffered, and never learned." That phrase recalled an attack he made on the press long ago, in the days of Watergate, when he accused the Washington Post of being in bed with George McGovern. "There is a cultural and social affinity between the McGovernites and the Post executives and editors," Dole said then. "They belong to the same elite: They can be found living cheek-by-jowl in the same exclusive chic neighborhoods, and hob-nobbing at the same Georgetown parties." The deeper story here isn't whether Dole was wrongly shunted onto D19 when he ought to have been on A1. It's his feelings, as he says goodbye to politics, about the people who get to decide. | C. they had no full-time reporters following him |
What is the initial imaging finding suggestive of tuberculosis in Mr. Hurley?
Choose the correct answer from the following options:
A. Splenomegaly
B. Mediastinal lymph node conglomerate
C. Liver tuberculosis
D. Bronchial and vascular stenosis
E. Neck/Thorax CT showing regression of pulmonary infiltrates
| ### Patient Report 0
**Dear colleague, **
We are reporting on Mr. Bruno Hurley, born on 12/24/1965, who has been
under our outpatient treatment since 02/11/2020.
**Diagnoses:**
- Refractory tuberculosis
- Manifestations: Open pulmonary tuberculosis, lymph node tuberculosis
(cervical, hilar, mediastinal), liver tuberculosis
**Imaging:**
- 11/01/19 Chest CT: Mediastinal lymph node conglomerate centrally
with poststenotic infiltrates on both sides. Splenomegaly.
- 11/04/19 Bronchoscopy: Large mediastinal and right hilar lymphomas.
Subcritical constriction of right segmental bronchi. EBUS-TBNA LK4R
and 10/11R.
**Microbiology:**
- 11/04/19 Tracheobronchial Secretions: Microscopic detection of
acid-fast rods, cultural detection of Mycobacterium tuberculosis,
phenotypically no evidence of resistance.
**Therapy:**
- Initial omission of pyrazinamide due to pancytopenia.
- Moxifloxacin: 11/10/19-11/20/19
- Pyrazinamide: 11/20/19-02/11/20
- Ethambutol: 11/08/19-02/11/20
- Rifampicin since: 11/08/19
- Isoniazid since: 11/08/19
- Levofloxacin since: 02/11/20
- Immunomodulatory therapy for low basal interferon / interferon
levels (ACTIMMUNE®)
**Microbiology:**
- 01/20/20 Sputum: Cultural detection of Mycobacterium tuberculosis:
Phenotypically no evidence of resistance.
- 01/02/20 Sputum: Last cultural detection of Mycobacterium
tuberculosis.
- 06/15/20 BAL: Occasional acid-fast rods, 16S-rRNA-PCR: M.
tuberculosis complex, no cultural evidence of Mycobacteria.
- 06/15/20 Lung biopsy: Occasional acid-fast rods, no cultural
evidence of Mycobacteria.
- 03/12/21 Sputum: first sputum without acid-fast rods, consistently
microscopically negative sputum samples since then.
**Histology:**
- 07/16/21: Mediastinal lymph node biopsy: Histologically no evidence
of malignancy/lymphoma.
**Other Diagnoses: **
- Secondary Acute Myeloid Leukemia with Myelodysplastic Syndrome
- Blood count at initial diagnosis: 15% blasts, erythrocyte
substitution required.
**Therapy:**
- 12/20-03/21 TB therapy
- 02/20-01/21 TB therapy: RMP + INH + FQ
- 01/21-04/21 RMP + INH + FQ + Actimmune® 04/22 CT: Regressive
findings of pulmonary TB changes, regressive cervical lymph nodes,
mediastinal LAP, and liver lesions size-stable; Sputum: No acid-fast
rods detected for the first time since 03/21.
- BM aspiration: Secondary AML.
**Current Presentation:** Admission for allogeneic stem cell
transplantation
Pathogen Location / Material of Detection or Infection Month/Year or
Last Detection
- HIV Serology: Negative - 11/19
- Mycobacterium tuberculosis Complex: Bronchoalveolar Lavage,
Tracheobronchial Secretion, Sputum - 11/19
**Medical History:** We took over Mr. Hurley for the continuation of TB
therapy on 11/02/20. His hospital admission took place at the end of
October 2019 due to neutropenic fever. The patient reported temperatures
up to 39°C for the past 3 days. Since 08/19, the patient has been
receiving hematological-oncological treatment for MDS. The colleagues
from hematology performed a repeat bone marrow aspiration before
transferring to Station 12. The blast percentage was significantly
reduced. HLA typing of the brother for allogeneic stem cell
transplantation planning had already been done in the summer of 2019.
After a chest CT revealed extensive mediastinal lymphomas with
compression of the bronchial tree bilaterally and post-stenotic
infiltrates, a bronchoscopy was performed. M. tuberculosis was cultured
from sputum and TBS. An EBUS-guided lymph node biopsy was histologically
processed, revealing granulomatous inflammation and molecular evidence
of the M. tuberculosis complex. On 11/08/19, a four-drug
anti-tuberculosis therapy was initiated, initially with Moxifloxacin
instead of Pyrazinamide due to pancytopenia. Moxifloxacin was replaced
by Pyrazinamide on 11/20/19. The four-drug therapy was continued for a
total of 3 months due to prolonged microscopic evidence of acid-fast
rods in follow-up sputum samples. Isoniazid dosage was adjusted after
peak level control (450 mg q24h), as was Rifampicin dose (900 mg q24h).
On 01/02/20, Mycobacterium tuberculosis was last cultured in a sputum
sample. Nevertheless, acid-fast rods continued to be detected in the
sputum. Due to the lack of culturability of mycobacteria, Mr. Hurley was
discharged to home care after consultation with the Tuberculosis Welfare
Office.
**Allergies**: None known. Toxic Substances: Smoking: Non-smoker;
Alcohol: No; Drugs: No
**Social History:** Originally from Brazil, has been living in the US
for 8 years. Lives with his partner.
**Current lab results:**
**Parameter** **Results** **Reference Range**
---------------------------------------- -------------- ---------------------
ConA-Induced Cytokines (Th1/Th2)
Interleukin 10 10 pg/mL \< 364 pg/mL
Interferon Gamma 265-6781 pg/mL
Interleukin 2 74 pg/mL 43-374 pg/mL
Interleukin 4 13 pg/mL \< 34 pg/mL
Interleukin 5 3 pg/mL \< 55 pg/mL
Naive CD45RA+CCR7+ (% of CD8+) 6.35 % 8.22-59.58 %
TEMRA CD45RA+CCR7- (% of CD8+) 50.44 % 7.32-55.99 %
Central Memory CD45RA-CCR7+ (% of CD) 2.60 % 1.67-5.84 %
Effector Memory CD45RA-CCR7- (% of CD) 40.60 % 22.52-62.25 %
Naive CD45RA+ (% of CD4+) 26.26 % 17.46-60.24 %
TEMRA CD45RA+ CCR7- (% of CD4+) 1.26 % 2.74-15.54 %
Central Memory CD45RA-CCR7+ (% of CD) 34.21 % 16.40-33.41 %
Effector Memory CD45RA-CCR7- (% of CD) 38.28 % 17.38-40.38 %
Granulocytes 0.60 abs./nL 3.00-6.50 abs./nL
Granulocytes (relative) 45 % 50-80 %
Lymphocytes 0.57 abs./nL 1.50-3.00 abs./nL
Lymphocytes (relative) 43 % 20-40 %
Monocytes 0.13 abs./nL \<0.50 abs./nL
Monocytes (relative) 10 % 2-10 %
NK Cells 0.16 abs./nL 0.10-0.40 abs./nL
NK Cells (% of Lymphocytes) 29 5-25
γ/δ TCR+ T-Cells (relative) 2 % \< 10 %
α/β TCR+ T-Cells (relative) 98 % \>90 %
CD19+ B-Cells (% of Lymphocytes) 3 % 5-25 %
CD4/CD8 Ratio 0.9 % 1.1-3.0 %
CD8-CD4-T-Cells (% of T-Cells) 5.86 % \< 15.00 %
CD8+CD4+-T-Cells (% of T-Cells) 0.74 % \< 10.00 %
CD3+ T-Cells 0.38 abs./nL 0.90-2.20 abs./nL
**Parameter** **Results** **Reference Range**
-------------------------------------------------- ---------------- ---------------------
Complete Blood Count (EDTA)
Hemoglobin 6.6 g/dL 13.5-17.0 g/dL
Hematocrit 19.0 % 39.5-50.5 %
Erythrocytes 2.3 x 10\^6/uL 4.3-5.8 x 10\^6/uL
Platelets 61 x 10\^3/uL 150-370 x 10\^3/uL
MCV (Mean Corpuscular Volume) 81.5 fL 80.0-99.0 fL
MCH (Mean Corpuscular Hemoglobin) 28.3 pg 27.0-33.5 pg
MCHC (Mean Corpuscular Hemoglobin Concentration) 34.7 g/dL 31.5-36.0 g/dL
MPV (Mean Platelet Volume) 10.4 fL 7.0-12.0 fL
RDW-CV (Red Cell Distribution Width-CV) 12.7 % 11.5-15.0 %
**Parameter** **Results** **Reference Range**
-------------------------- ------------- ---------------------
Other Investigations
QFT-TB Gold plus TB1 0.11 IU/mL \<0.35 IU/mL
QFT-TB Gold plus TB2 0.07 IU/mL \<0.35 IU/mL
QFT-TB Gold plus Mitogen 3.38 IU/mL \>0.50 IU/mL
QFT-TB Gold plus Result Negative
**Lung Aspiration from 06/15/20:** Examination Request: Acid-fast rods
(Microscopy + Culture) **Microscopic Findings:**
- Auramine stain: Occasionally, acid-fast rods Result: No growth of
Mycobacterium sp. after 12 weeks of incubation.
2. Forceps Biopsy Exophytic Trachea: One piece of tissue. Microscopy:
HE, PAS, Giemsa, Diagnosis:
3. Predominantly blood clot and necrotic material alongside sparsely
altered lymphatic tissue due to sampling (EBUS-TBNA LK 7 as
indicated).
4. Components of a granulation tissue polyp (Forceps Biopsy Exophytic
Trachea as indicated). Comment: The finding in 1. continues to be
suspicious of a mycobacterial infection. We are conducting molecular
pathological examinations in this regard and will report again.
> [Comment]{.underline}: Detection of mycobacterial DNA of the M.
> tuberculosis complex type. No evidence of atypical mycobacteria. No
> evidence of malignancy.
**Current Medication:**
**Medication** **Dosage** **Frequency**
------------------------- ------------ ---------------
Rifampin (Rifadin) 600 mg 1-0-0
Isoniazid (Nydrazid) 500 mg 1-0-0
Levofloxacin (Levaquin) 450 mg 1-0-1
**\
**
### Patient Report 1
**Dear colleague, **
We report to you about our patient Mr. Bruno Hurley, born on 12/24/1965.
Who has received inpatient treatment from 07/17/2021 to 09/03/2021.
**Diagnoses**:
- Acute Myeloid Leukemia with Myelodysplasia-Related Changes (AML-MRC)
<!-- -->
- Myelodysplastic Syndrome EB-2, diagnosed in July 2010. Blood count
at initial diagnosis: 15% blasts, erythrocyte transfusion-dependent.
Cytogenetics: 46,XY \[1\]; 47,XY,+Y,i(21)(q10)\[15\];
47,XY,+Y,trp(21)(q11q22)\[4\]. Molecular genetics: Mutations in
RUNX1, SF3B1. IPSS-R: 7 (very high risk).
- In 08/2020, diagnosed with Myelodysplastic Syndrome with ring
sideroblasts.
- Received transfusions of 2 units of red blood cells every 3-4 weeks
to maintain hemoglobin between 4-6 g/dL.
- Bone marrow biopsy showed MDS-EB2 with 14.5% blasts.
- Initiated Azacitidine treatment (2x 75 mg subcutaneously, days 1-5 +
8-9 every 4 weeks) as an outpatient.
- 10/23/2019: Hospitalized for fever during neutropenia.
- 12/06/2019: Diagnosed with tuberculosis - positive Tbc-PCR in
tracheobronchial secretions, acid-fast bacilli in tracheobronchial
secretions, histological confirmation from EBUS biopsy of a
conglomerate of melted lymph nodes from 11/03/2019.
- 01/2021: Bone marrow biopsy showed secondary AML with 26% blasts.
- 03/2021: Started Venetoclax/Vidaza.
- 05/2021: Bone marrow biopsy showed 0.8% myeloid blasts coexpressing
CD117 and CD7. Cytology showed 6% blasts.
- 05/2021: Started the 4th cycle of Vidaza/Venetoclax.
- 06/17/2022: Started the 5th cycle of Vidaza/Venetoclax.
- 07/29/2021: Underwent allogeneic stem cell transplantation from a
HLA-identical unrelated donor (10/10 antigen match) for AML-MRC in
first complete remission (CR). Conditioning regimen included
Treosulfan 12g/m2, Fludarabin 5x 30 mg/m2, ATG 3x 10 mg/kg.
**Other Diagnoses:**
- Persistent tuberculosis with lymph node swelling since June 2020.
- Open lung tuberculosis diagnosed in November 2019.
- Location: CT of the chest showed central mediastinal lymph node
conglomerate with post-stenotic infiltrates bilaterally,
splenomegaly.
- Bronchoscopy on December 5, 2020, showed large mediastinal and
right hilar lymph nodes, subcritical narrowing of right
segmental bronchi. EBUS-TBNA
- CT Chest/Neck on 02/05/2020: Regression of pulmonary
infiltrates, enlargement of necrotic lymph nodes in the upper
mediastinum and infraclavicular on the right (compressing the
internal jugular vein/esophagus).
- Culture confirmation of Mycobacterium tuberculosis,
pansensitive: Tracheobronchial secretion
- Initiated antituberculous combination therapy
**Current Presentation:** Admission for allogeneic stem cell
transplantation from a HLA-identical unrelated donor (10/10 antigen
match) for AML-MRC in first complete remission.
**Medical History:** In 2019, Mr. Hurley was diagnosed with
Myelodysplastic Syndrome EB-2. Starting from September 2019, he received
Azacitidine therapy. In December 2020, he was diagnosed with open lung
tuberculosis, which was challenging to treat due to his dysfunctional
immune system. In January 2021, his MDS progressed to AML-MRC with 26%
blasts. After treatment with Venetoclax/Vidaza, he achieved remission in
May 2021. Tuberculosis remained largely under control.
Due to AML-MRC, he was recommended for allogeneic stem cell
transplantation from a HLA-identical unrelated donor. At the time of
admission for transplantation, he was largely asymptomatic. He
occasionally experienced mild dry cough but denied fever, night sweats,
or weight loss. The admission and counseling were conducted with
translation assistance from his life partner due to limited proficiency
in English.
**Allergies**: None
**Transfusion History**: Currently requires transfusions every 14 days.
Both red blood cell and platelet transfusions have been tolerated
without problems.
**Abdominal CT from 01/20/2021:**
**Findings**: Significant peripancreatic fluid accumulation in the upper
abdominal area with a somewhat indistinct border between the pancreatic
tissue, particularly in the pancreatic head region. Evidence of
inflammation affecting the stomach and duodenum. No presence of free air
or indications of hollow organ perforation. No conclusive signs of a
well-defined abscess. Moreover, the other parenchymal abdominal organs,
especially those lacking focal abnormalities suggestive of neoplastic or
inflammatory conditions, displayed normal appearances. The gallbladder
showed no notable issues, and there were no radiopaque concretions
observed. Both the intra- and extrahepatic bile ducts appeared
adequately dilated. Abdominal hollow organs exhibited unremarkable and
normal appearances without corresponding contrast and dilation. The
appendix appeared within normal parameters. Abdominal lymph nodes showed
no unusual findings. Some degree of aortic vasosclerosis was noted. The
depiction of the included lung portions revealed no abnormalities.
**Results**: Findings indicative of acute pancreatitis, most likely of
an exudative nature. No signs of hollow organ perforation were detected,
and there was no definitive evidence of an abscess (as far as could be
determined from native imaging).
**Summary**: The patient was admitted to our hospital through the
emergency department with the symptoms described above. With typical
upper abdominal pain and significantly elevated serum lipase levels, we
diagnosed acute pancreatitis. This diagnosis was corroborated by
peripancreatic fluid and ill-defined organ involvement in the abdominal
CT scan. There were no laboratory or anamnestic indications of a biliary
origin. The patient denied excessive alcohol consumption.
**Medication upon Admission:**
**Medication** **Dosage** **Frequency**
------------------------- ------------ ---------------
Rifampin (Rifadin) 600 mg 1-0-0
Isoniazid (Nydrazid) 500 mg 1-0-0
Levofloxacin (Levaquin) 450 mg 1-0-1
**Physical Examination:** General: Oriented in all qualities, in good
general condition with normal body weight (75 kg, 187 cm) Vital signs at
admission: Heart rate 63/min, Blood pressure 110/78 mmHg. Temperature at
admission 36.8 °C, Oxygen saturation 100% on room air. Skin and mucous
membranes: Dry skin, normal skin color, normal skin turgor. No scleral
icterus, non-irritated conjunctiva. Normal oral mucosa, moist tongue
without coating, no ulcers or thrush. Heart: Normal heart sounds,
rhythmic, regular rate, no pathological heart murmurs heard on
auscultation. Lungs: Resonant percussion sound, clear breath sounds
bilaterally, no wheezing, no prolonged expiration. Abdomen: Unremarkable
scar tissue, normal bowel sounds in all quadrants, soft, non-tender, no
guarding, liver and spleen not enlarged. Vascular: Central and
peripheral pulses palpable, no jugular vein distention, no peripheral
edema, extremities warm with no significant difference in size. Lymph
nodes: Palpable cervical swelling, inguinal and axillary lymph nodes
unremarkable. Neurology: Grossly neurologically unremarkable.
On 08/22/2021, a four-lumen central venous catheter was placed in the
right internal jugular vein without complications. During the
conditioning regimen, the patient received the following:
**Medication** **Dosage** **Frequency**
---------------------------------------------- ---------------------- -------------------------
Fludarabine (Fludara) 30 mg/m² (5x 57 mg) 07/23/2023 - 07/27/2023
Treosulfan (Ovastat) 12 g/m² (3x 22.9 g) 07/23/2023 - 07/25/2023
Anti-Thymocyte Globulin (ATG, Thymoglobulin) 10 mg/kg (3x 700 mg) 07/23/2023 - 07/28/2023
**Antiemetic Therapy:**
The antiemetic therapy included Ondansetron, Aprepitant, and
Dexamethasone, and the conditioning regimen was well tolerated.
**Prophylaxis of Graft-Versus-Host Disease (GvHD):**
**Substances** **Start Date** **Day -2** **Day 1**
---------------- ---------------- ------------ -----------
Cyclosporine 08/28/2022
Mycophenolate 07/30/2021
**Stem Cell Source** **Date** **CD34/kg KG** **CD45/kg KG** **CD3/kg KG** **Volume**
---------------------- ------------ ---------------- ---------------- --------------- ------------
PBSCT 07/29/2021 7.39 x10\^6 8.56 x10\^8 260.7 x10\^6 194 ml
**Summary:** Mr. Hurley was admitted for allogeneic stem cell
transplantation from a HLA-identical unrelated donor for AML-MRC. The
conditioning regimen with Treosulfan, Fludarabin, and ATG was well
tolerated, and the transplantation proceeded without complications.
**Toxicities:** There was an adverse event-related increase in bilirubin
levels, reaching a maximum of 2.68 mg/dL. Elevated ALT levels, up to a
maximum of 53 U/L, were observed.
**Acute Graft-Versus-Host Disease (GvHD):** Signs of GvHD were not
observed until the time of discharge.
**Medication upon Discharge:**Formularbeginn
**Medication ** **Dosage** **Frequency**
---------------------------------- ------------ ------------------------------------------------
Acyclovir (Zovirax) 500 mg 1-0-1-0
Entecavir (Baraclude) 0.5 mg 1-0-0-0
Rifampin (Rifadin) 600 mg 1.5-0-0-0
Isoniazid/Pyridoxine (Nydrazid) 300 mg 2-0-0-0
Levofloxacin (Levaquin) 500 mg 1-0-1-0
Mycophenolate Mofetil (CellCept) 500 mg 2-0-2-0
Folic Acid 5 mg 1-0-0-0
Magnesium \-- 3-3-3-0
Pantoprazole (Protonix) 40 mg 1-0-0-0 (before a meal)
Ursodeoxycholic Acid (Actigall) 250 mg 1-1-1-0
Cyclosporine (Sandimmune) 100 mg 100 mg 4-0-4-0
Cyclosporine (Sandimmune) 50 mg 50 mg 4-0-4-0 (based on TDM, last dose 400 mg 1-0-1)
Cyclosporine (Sandimmune) 10 mg 10 mg 4-0-4-0 (based on TDM, last dose 400 mg 1-0-1)
**Current Recommendations: **
1. Bone marrow puncture on Day +60, +120, and +360 post-transplantation
(including MRD and chimerism) and Day +180 depending on MRD and
chimerism progression.
2. Continuation of immunosuppressive therapy with ciclosporin adjusted
to achieve target levels of around 150 ng/ml, for a minimum of 3
months post-transplantation. Immunosuppression with mycophenolate
mofetil will be continued until Day +40.
3. Prophylaxis with Aciclovir must continue for 6 weeks after
discontinuation of immunosuppression at a dosage of 15-20 mg/kg/day
(divided into 2 doses). Dose adjustment based on renal function may
be necessary.
4. Pneumocystis pneumonia prophylaxis through monthly Pentamidine
inhalation or administration of Cotrim forte 960mg must continue at
least until immunosuppression is discontinued or until an absolute
CD4+ T-cell count exceeds \>200/µL in peripheral blood. Cotrim forte
960mg has not been started when leukocytes are \<2/nL.
5. Weekly monitoring of CMV and EBV viral loads through quantitative
PCR from EDTA blood.
6. Timing of antituberculous medication intake:
- Take Rifampicin and Isoniazid in the morning on an empty stomach, 30
minutes before breakfast.
- Take levofloxacin with a 2-hour gap from divalent cations (Mg2+,
strongly calcium-rich foods).
**Lab results upon Discharge:**
**Parameter** **Result** **Reference Range**
------------------------------ -------------------- ---------------------
Cyclosporine 127.00 ng/mL \--
Sodium 141 mEq/L 136-145 mEq/L
Potassium 4.1 mEq/L 3.5-4.5 mEq/L
Glucose 108 mg/dL 60-110 mg/dL
Creatinine 0.65 mg/dL 0.70-1.20 mg/dL
Estimated GFR (eGFR CKD-EPI) 111 mL/min/1.73 m² \--
Urea 26 mg/dL 17-48 mg/dL
Total Bilirubin 0.35 mg/dL \<1.20 mg/dL
**Complete Blood Count **
**Parameter** **Results** **Reference Range**
--------------- ----------------- -----------------------
Hemoglobin 9.5 g/dL 13.5-17.0 g/dL
Hematocrit 28.2% 39.5-50.5%
Erythrocytes 3.2 x 10\^6/µL 4.3-5.8 x 10\^6/µL
Leukocytes 1.47 x 10\^3/µL 3.90-10.50 x 10\^3/µL
Platelets 193 x 10\^3/µL 150-370 x 10\^3/µL
MCV 88.7 fL 80.0-99.0 fL
MCH 29.9 pg 27.0-33.5 pg
MCHC 33.7 g/dL 31.5-36.0 g/dL
MPV 9.8 fL 7.0-12.0 fL
RDW-CV 18.9% 11.5-15.0%
### Patient Report 2
**Dear colleague, **
We report on Mr. Bruno Hurley, born on 12/24/1965, who was under our
inpatient care from 2/20/2022, to 02/24/2022.
**Diagnoses:**
- Acute Pancreatitis, possibly medication-related under antitubercular
therapy.
- Current medications include Entecavir, Rifampicin, and
Isoniazid/Pyridoxin, which have been paused after consultation with
the infectious disease team.
**Other Diagnoses:**
- Acute Myeloid Leukemia with Myelodysplasia-Related Changes (AML-MRC)
- Myelodysplastic Syndrome EB-2
- Allogeneic stem cell transplantation
- EBV Reactivation (Treated with immunoglobulins for 3 days)
- Persistent Tuberculosis with lymph node swelling
- Open Lung Tuberculosis - Initial Diagnosis
- Antitubercular combination therapy since (Moxifloxacin, Pyrazinamid,
Ethambutol, Rifampicin, Isoniazid).
- Rectal colonization with 4-MRGN.
**Medical History:** The patient presented via ambulance from his
workplace. The patient reported sudden onset upper abdominal pain,
mainly in the epigastric region, accompanied by nausea and vomiting. He
also experienced watery diarrhea once today. He had lunch around noon,
consisting noodles. There was no fever, cough, sputum production,
dyspnea, or urinary abnormalities. He has been taking daily
antitubercular combination therapy, including Rifampicin, for open
tuberculosis. The patient denied alcohol consumption and weight loss.
**Medication** **Dosage** **Frequency**
----------------------------------------- ------------ ----------------------------
Acyclovir (Zovirax) 400 mg 1-0-1
Entecavir (Baraclude) 0.5 mg 1-0-0
Rifampin (Rifadin) 600 mg 1.5-0-0
Isoniazid/Pyridoxine (Nydrazid) 300 mg 1-0-1
Pantoprazole (Protonix) 40 mg 1-0-0
Trimethoprim/Sulfamethoxazole (Bactrim) 960 mg 1 tablet, on Mon, Wed, Fri
Methylprednisolone (Medrol) 0.79 mg As needed
Prednisolone 4 mg As needed
**Allergies:** None
**Physical Exam:**
Vital Signs: Blood Pressure 178/90 mmHg, Pulse 85/min, SpO2 89%,
Temperature 36.7°C, Respiratory Rate 20/min.
Clinical Status: Upon initial examination, a reduced general condition.
Cardiovascular: Heart sounds were normal, rhythm was regular, and no
murmurs were heard.
Respiratory: Vesicular breath sounds, sonorous percussion.
Abdominal: Sluggish peristalsis, soft abdominal walls, guarding and
tenderness in the epigastrium, liver and spleen not palpable, no free
fluid.
Extremities: Minimal edema.
**ECG Findings:** ECG on admission showed normal sinus rhythm (69/min),
normal ST intervals, R/S transition in V3/V4, and no significant
abnormalities.
´
**Medication upon Discharge:**
**Medication ** **Dosage** **Frequency**
----------------------------------------- ------------ ----------------------------
Acyclovir (Zovirax) 400 mg 1-0-1
Entecavir (Baraclude) 0.5 mg PAUSED
Rifampin (Rifadin) 600 mg PAUSED
Isoniazid/Pyridoxine (Nydrazid) 300 mg PAUSED
Pantoprazole (Protonix) 40 mg 1-0-0
Trimethoprim/Sulfamethoxazole (Bactrim) 960 mg 1 tablet, on Mon, Wed, Fri
Methylprednisolone (Medrol) 0.79 mg As needed (as needed)
Prednisolone 4 mg As needed (as needed)
Tramadol (Ultram) 50 mg 1 tablet, every 6 hours
**Parameter** **Results** **Reference Range**
------------------------- ----------------- -----------------------
White Blood Cells (WBC) 5.0 x 10\^9/L 3.7 - 9.9 x 10\^9/L
Hemoglobin 14.0 g/dL 13.6 - 17.5 g/dL
Hematocrit 40% 40 - 53%
Red Blood Cells (RBC) 4.00 x 10\^12/L 4.4 - 5.9 x 10\^12/L
MCV 99 fL 80 - 96 fL
MCH 32.8 pg 28.3 - 33.5 pg
MCHC 33.1 g/dL 31.5 - 34.5 g/dL
Platelets 161 x 10\^9/L 146 - 328 x 10\^9/L
Absolute Neutrophils 3.7 x 10\^9/L 1.8 - 6.2 x 10\^9/L
Absolute Monocytes 0.31 x 10\^9/L 0.25 - 0.85 x 10\^9/L
Absolute Eosinophils 0.03 x 10\^9/L 0.03 - 0.44 x 10\^9/L
Absolute Basophils 0.01 x 10\^9/L 0.01 - 0.08 x 10\^9/L
Absolute Lymphocytes 0.9 x 10\^9/L 1.1 - 3.2 x 10\^9/L
Immature Granulocytes 0.0 x 10\^9/L 0.0 x 10\^9/L
### Patient Report 3
**Dear colleague, **
We are writing to inform you on our patient, Mr. Hurley, who presented
to our outpatient clinic on 07/12/2022.
**Diagnoses:**
- Acute Pancreatitis, possibly medication-related under antitubercular
therapy.
- Current medications include Entecavir, Rifampicin, and
Isoniazid/Pyridoxin, which have been paused after consultation with
the infectious disease team.
**Other Diagnoses:**
- Acute Myeloid Leukemia with Myelodysplasia-Related Changes (AML-MRC)
- Myelodysplastic Syndrome EB-2
- Allogeneic stem cell transplantation
- EBV Reactivation (Treated with immunoglobulins for 3 days)
- Persistent Tuberculosis with lymph node swelling
- Open Lung Tuberculosis - Initial Diagnosis
- Antitubercular combination therapy since (Moxifloxacin, Pyrazinamid,
Ethambutol, Rifampicin, Isoniazid).
- Rectal colonization with 4-MRGN.
**Current Presentation:** Presented with a referral from outpatient
oncologist for suspected recurrent AML, with DD GvHD ITP in the setting
of progressive pancytopenia, primarily thrombocytopenia. The patient is
in good general condition, denying acute symptoms, particularly no rash,
diarrhea, dyspnea, or fever.
**Physical Examination:** Alert, oriented, no signs of respiratory
distress, heart sounds regular, abdomen soft, no tenderness, no skin
rashes, especially no signs of GvHD, no edema.
- Heart Rate (HR): 130/85
- Temperature (Temp): 36.7°C
- Oxygen Saturation (SpO2): 97
- Respiratory Rate (AF): 12
- Pupillary Response: 15
**Imaging (CT):**
- [11/04/19 CT Chest/Abdomen/Pelvis ]{.underline}
- [01/04/20 Chest CT:]{.underline} Marked necrotic lymph nodes hilar
right with bronchus and vascular stenosis. Significant increasing
pneumonic infiltrates predominantly on the right.
- [02/05/20 Neck/Chest CT]{.underline}: Regression of pulmonary
infiltrates, but increased size of necrotic lymph nodes, especially
in the upper mediastinum and right infraclavicular with slit-like
compression of the right internal jugular vein and the esophagus.
- [06/07/20 Neck/Chest CT]{.underline}: Size-stable necrotic lymph
node conglomerate infraclavicular right, dimensioned axially up to
about 6 x 2 cm, with ongoing slit-shaped compression of the right
internal jugular vein. Hypoplastic mastoid cells left, idem.
Progressive, partly new and large-volume consolidations with
adjacent ground glass infiltrates on the right in the anterior, less
posterior upper lobe and perihilar. Inhomogeneous, partially reduced
contrast of consolidated lung parenchyma, broncho pneumogram
preserved dorsally only.
- [10/02/20 Neck/Chest/Abdomen/Pelvis CT:]{.underline} Size-regressive
consolidating infiltrate in the right upper lobe and adjacent
central lower lobe with increasing signs of liquefaction.
Progressive right pleural effusion and progressive signs of
pulmonary volume load. Regressive cervical, mediastinal, and right
hilar lymphadenopathy. Ongoing central hilar conglomerates that
compress the central hilar structures. Partly constant, partly
regressive presentation of known tuberculosis-suspected liver
lesions.
- [12/02/20 CT Chest/Abdomen/Pelvis]{.underline}.
- [01/20/2021 Abdominal CT.]{.underline}
- [02/23/21 Neck/Chest CT:]{.underline} Slightly regressive/nodular
fibrosing infiltrate in the right upper lobe and adjacent central
lower lobe with continuing significant residual findings. Within the
infiltrate, larger poorly perfused areas with cavitations and
scarred bronchiectasis. Increasing, partly patchy densities on the
left basal region, differential diagnoses include infiltrates and
ventilation disorders. Essentially constant cervical, mediastinal,
and hilar lymphadenopathy. Constant liver lesions in the upper
abdomen, differential diagnoses include TB manifestations and cystic
changes.
- [06/12/21 Neck/Chest/Abdomen/Pelvis CT:]{.underline} Improved
ventilation with regressive necrotic TB manifestations perihilar,
now only subtotal lobar atelectasis. Essentially constant necrotic
lymph node manifestations cervical, mediastinal, and right hilar,
exemplarily suprasternal right or right paratracheal. Narrow right
pleural effusion. Medium-term constant hypodense liver lesions
(regressive).
**Patient History:** Known to have AML with myelodysplastic changes,
first diagnosed 01/2021, myelodysplastic syndrome EB-2, fist diagnoses
07/2019, and history of allogeneic stem cell transplantation.
**Treatment and Progression:** Patient is hemodynamically stable, vital
signs within normal limits, afebrile. In good general condition,
clinical examination unremarkable, especially no skin GvHD signs. Venous
blood gas: Acid-base status balanced, electrolytes within normal range.
Laboratory findings show pancytopenia, Hb 11.3 g/dL, thrombocytopenia
29/nL, leukopenia 1.8/nL, atypical lymphocytes described as \"resembling
CLL,\" no blasts noted. Consultation with Hemato/Oncology confirmed no
acute need for hospitalization. Follow-up in the Hemato-Oncological
Clinic in September.
**Imaging:**
**CT Chest/Abdomen/Pelvis on 11/04/19:**
**Assessment:** In comparison with 10/23/19: In today\'s
contrast-enhanced examination, a newly unmasked large tumor is noted in
the right pulmonary hilum with encasement of the conduits of the right
lung lobe. Differential diagnosis includes a lymph node conglomerate,
central bronchial carcinoma, or, less likely, an inflammatory lesion.
Multiple suspicious malignant enlarged mediastinal lymph nodes,
particularly on the right paratracheal and infracarinal regions.
Short-term progression of peribronchovascular consolidation in the right
upper lobe and multiple new subsolid micronodules bilaterally.
Differential diagnosis includes inflammatory lesions, especially in the
presence of known neutropenia, which could raise suspicion of fungal
infection.
Intraabdominally, there is an image suggestive of small bowel subileus
without a clearly defined mechanical obstruction.
Density-elevated and ill-defined cystic lesion in the left upper pole of
the kidney. Primary consideration is a hemorrhaged or thickened cyst,
but ultimately, the nature of the lesion remains uncertain.
**CT Chest on 01/04/20:** Significant necrotic hilar lymph nodes on the
right with bronchial and vascular stenosis. Marked progression of
pulmonary infiltrates, particularly on the right, still compatible with
superinfection in the context of known active tuberculosis
**CT Chest from 02/05/20**: Marked necrotic lymph nodes hilar right with
bronchial and vascular stenosis. Significantly increasing pneumonia-like
infiltrates, particularly on the right, still compatible with
superinfection in the context of known open tuberculosis.
**Neck/Chest CT from 06/07/20:** Size-stable necrotic lymph node
conglomerate infraclavicular right, dimensioned axially up to about 6 x
2 cm, with ongoing slit-shaped compression of the right internal jugular
vein. Hypoplastic mastoid cells left, idem. Progressive, partly new and
large-volume consolidations with adjacent ground glass infiltrates on
the right in the anterior, less posterior upper lobe and perihilar.
Inhomogeneous, partially reduced contrast of consolidated lung
parenchyma, broncho pneumogram preserved dorsally only.
**Neck Ultrasound from 08/14/2020:** Clinical History, Question,
Justifying Indication: Follow-up of cervical lymph nodes in
tuberculosis.
**Findings/Assessment:** Neck Lymph Node Ultrasound from 05/20/2020 for
comparison. As in the previous examination, evidence of two
significantly enlarged supraclavicular lymph nodes on the right, both
showing a decrease in size compared to the previous examination: The
more medial node measures 2.9 x 1.6 cm compared to the previous 3.7 x
1.7 cm, while the more laterally located lymph node measures 3.3 x 1.4
cm compared to the previous 4.2 x 1.5 cm. The more medial lymph node
appears centrally hypoechoic, indicative of partial liquefaction, while
the more lateral lymph node has a rather solid appearance. No other
pathologically enlarged lymph nodes detected in the cervical region.
**CT Neck/Chest/Abdomen/Pelvis from 10/02/2020:** Assessment: Compared
to the previous examination from 06/07/2020, there is evidence of
regression in findings: Size regression of consolidating infiltrate in
the right upper lobe and the adjacent central lower lobe, albeit with
increasing signs of cavitation. Progressive right pleural effusion and
progressive signs of pulmonary volume overload. Regression of cervical,
mediastinal, and right hilar lymphadenopathy. Persistent centrally
liquefying lymph node conglomerates in the right hilar region,
compressing central hilar structures. Some findings remain stable, while
others have regressed. No evidence of new manifestations.
**CT Chest/Abdomen/Pelvis from 12/02/20:** Assessment: Compared to
10/02/20: In today\'s contrast examination, a newly unmasked large tumor
is located right pulmonary hilar, encasing the conduits of the right
lung lobe; Differential diagnosis includes lymph node conglomerate,
central bronchial carcinoma, and a distant possibility of inflammatory
lesions. Multiple suspiciously enlarged mediastinal lymph nodes,
especially right paratracheal and infracarinal. In a short time,
progressive peribronchovascular consolidations in the right upper lobe
and multiple new subsolid micronodules bilaterally; Differential
diagnosis includes inflammatory lesions, potentially fungal in the
context of known neutropenia. Intra-abdominally, there is a picture of
small bowel subileus without discernible mechanical obstruction.
Corresponding symptoms? Densely elevated and ill-defined cystic lesion
in the upper pole of the left kidney; Differential diagnosis primarily
includes a hemorrhaged/thickened cyst, ultimately with uncertain
malignancy.
**Chest in two planes on 04/23/2021:** **Findings/Assessment:** In
comparison with the corresponding prior images, most recently on
08/14/2020. Also refer to CT Neck and Chest on 01/23/2021. The heart is
enlarged with a leftward emphasis, but there are no signs of acute
congestion. Extensive consolidation projecting onto the right mid-field,
with a long-term trend toward regression but still clearly demarcated.
No pneumothorax. No pleural effusion. Known lymph nodes in the
mediastinum/hilum. Degenerative spinal changes.
**Neck/Chest CT on 02/23/21:** Slightly regressive/nodular fibrosing
infiltrate in the right upper lobe and adjacent central lower lobe with
continuing significant residual findings. Within the infiltrate, larger
poorly perfused areas with cavitations and scarred bronchiectasis.
Increasing, partly patchy densities on the left basal region,
differential diagnoses include infiltrates and ventilation disorders.
Essentially constant cervical, mediastinal, and hilar lymphadenopathy.
Constant liver lesions in the upper abdomen, differential diagnoses
include TB manifestations and cystic changes.
**CT Neck/Chest/Abdomen/Pelvis from 06/12/2021**: CT from 02/23/2021
available for comparison. Neck/Chest: Improved right upper lobe (ROL)
ventilation with regressive necrotic TB manifestations peri-hilar, now
only with subtotal lobar atelectasis. Essentially stable necrotic lymph
node manifestations in the cervical, mediastinal, and right hilar
regions, for example, supraclavicular on the right (18 mm, previously
30.1 Im 21.2) or right paratracheal (18 mm, previously 30.1 Im 33.8).
Narrow right pleural effusion, same as before. No pneumothorax. Heart
size normal. No pericardial effusion. Abdomen: Mid-term stable hypodense
liver lesions (regressing since 07/2021).
**Treatment and Progression:** Due to the extensive findings and the
untreatable immunocompromising underlying condition, we decided to
switch from a four-drug TB therapy to a three-drug therapy after nearly
3 months. In addition to rifampicin and isoniazid, levofloxacin was
initiated. Despite very good therapy adherence, acid-fast bacilli
continued to be detected microscopically in sputum samples without
culture confirmation of mycobacteria, even after discharge. Furthermore,
the radiological findings worsened. In April 2020, liver lesions were
identified in the CT that had not been described up to that point, and
pulmonary and mediastinal changes increased. Clinically, right cervical
lymphadenopathy also progressed in size. Due to a possible immune
reaction, a therapy with prednisolone was attempted for several weeks,
which did not lead to improvement. In June 2020, Mr. Hurley was
readmitted for bronchoscopy with BAL and EBUS-guided biopsy to rule out
differential diagnoses. An NTM-NGS-PCR was performed on the BAL, which
did not detect DNA from nontuberculous mycobacteria. Histologically,
predominantly necrotic material was found in the lymph node tissue, and
molecular pathological analysis detected DNA from the M. tuberculosis
complex. There were no indications of malignancy. In addition,
whole-genome sequencing of the most recently cultured mycobacteria was
performed, and latent resistance genes were also ruled out. Other
pathogens, including fungi, were likewise not detected. Aspergillus
antigen in BAL and serum was also negative. We continued the three-drug
therapy with Rifampicin, Isoniazid, and Levofloxacin. Mr. Hurley
developed an increasing need for red blood cell transfusions due to
myelodysplastic syndrome and began receiving regular transfusions from
his outpatient hematologist-oncologist in the summer of 2020. In a
repeat CT control in October 2020, increasing necrotic breakdown of the
right upper and middle lobes was observed, as well as progressive
ipsilateral pleural effusion and persistent mediastinal lymphadenopathy
and liver lesions. Mr. Hurley was referred to the immunology colleagues
to discuss additional immunological treatment options. After extensive
immune deficiency assessment, a low basal interferon-gamma level was
noted in the setting of lymphopenia due to MDS. In an immunological
conference, the patient was thoroughly discussed, and a trial of
interferon-gamma therapy in addition to antituberculous therapy was
discussed due to a low basal interferon-gamma level and a negative
Quantiferon test. After approval of an off-label application, we began
Actimmune® injections in January 2021 after extensive patient education.
Mr. Hurley learned to self-administer the subcutaneous injections and
initially tolerated the treatment well. Due to continuous worsening of
the blood count, a bone marrow puncture was performed again on an
outpatient basis by the attending hematologist-oncologist, and secondary
AML was diagnosed. Since February 2021, Mr. Hurley has received
Azacitidine and regular red blood cell and platelet concentrates. After
3 months of Actimmune® therapy, sputum no longer showed acid-fast
bacilli in March 2021, and radiologically, the left pleural effusion had
completely regressed, and the infiltrates had decreased. Actimmune® was
discontinued after 3 months. Towards the end of Actimmune® therapy, Mr.
Hurley developed pronounced shoulder arthralgia and pain in the upper
thoracic spine. Fractures were ruled out. With pain therapy, the pain
became tolerable and gradually improved. Arthralgia and myalgia are
common side effects of interferon-gamma. Due to the demonstrable
therapeutic response, we presented Mr. Hurley, along with an
interpreter, at the Department of Hematology and Oncology to discuss
further therapeutic options for AML in the context of the hematological
and infectious disease situation. After extensive explanation of the
disease situation, the risks of aggressive AML therapy in the presence
of unresolved tuberculosis, and the consequences of palliative AML
therapy. Mr. Hurley agreed to allogeneic stem cell transplantation after
some consideration. On an outpatient basis, the cytostatic therapy with
Azacitidine was expanded to include Venetoclax. Antituberculous therapy
with rifampicin, isoniazid, and levofloxacin was continued. Regular
sputum checks remained consistently microscopically negative until
complete AML remission was achieved. Mr. will be admitted for allogeneic
stem cell transplantation in July 2021. A repeat CT in June 2021
confirmed continued regression of the tuberculosis findings.
Antituberculous therapy will be continued indefinitely.
**CT Neck/Chest/Abdomen/Pelvis on 06/12/2022:** CT for comparison.
Neck/Chest: Improved right lung upper lobe ventilation with regressing
necrotic tuberculosis manifestations, now with only subtotal lobar
atelectasis. Essentially constant necrotic lymph node manifestations in
the cervical, mediastinal, and right hilar regions, as exemplified by
the right supraclavicular (18 mm, SE 301 HU 212) or right paratracheal
(18 mm, SE 301 HU 338) nodes. Narrow pleural effusion on the right,
likewise. No pneumothorax. The heart is not enlarged. No pericardial
effusion. Abdomen: Medium-term constant hypodense liver lesions
(regressing)
**Current Recommendations:** Continue antituberculous therapy without a
defined endpoint. Sputum checks during allogeneic stem cell
transplantation every 1-2 weeks. In case of clinical signs of persistent
infection, perform early CT scans of the neck, chest, and abdomen.
Follow-up appointment in our infectious diseases outpatient clinic after
allogeneic stem cell transplantation. | Mediastinal lymph node conglomerate |
Who is the man climbing the mountain?
A. A mountain guide looking for survivors
B. An astronomical surveyor who ended up there by accident
C. A mountaineer who happened to stumble upon an old radio
D. A Chang native looking for people on this planet
| ACCIDENTAL DEATH BY PETER BAILY The most dangerous of weapons is the one you don't know is loaded. Illustrated by Schoenherr The wind howled out of the northwest, blind with snow and barbed with ice crystals. All the way up the half-mile precipice it fingered and wrenched away at groaning ice-slabs. It screamed over the top, whirled snow in a dervish dance around the hollow there, piled snow into the long furrow plowed ruler-straight through streamlined hummocks of snow. The sun glinted on black rock glazed by ice, chasms and ridges and bridges of ice. It lit the snow slope to a frozen glare, penciled black shadow down the long furrow, and flashed at the furrow's end on a thing of metal and plastics, an artifact thrown down in the dead wilderness. Nothing grew, nothing flew, nothing walked, nothing talked. But the thing in the hollow was stirring in stiff jerks like a snake with its back broken or a clockwork toy running down. When the movements stopped, there was a click and a strange sound began. Thin, scratchy, inaudible more than a yard away, weary but still cocky, there leaked from the shape in the hollow the sound of a human voice. "I've tried my hands and arms and they seem to work," it began. "I've wiggled my toes with entire success. It's well on the cards that I'm all in one piece and not broken up at all, though I don't see how it could happen. Right now I don't feel like struggling up and finding out. I'm fine where I am. I'll just lie here for a while and relax, and get some of the story on tape. This suit's got a built-in recorder, I might as well use it. That way even if I'm not as well as I feel, I'll leave a message. You probably know we're back and wonder what went wrong. "I suppose I'm in a state of shock. That's why I can't seem to get up. Who wouldn't be shocked after luck like that? "I've always been lucky, I guess. Luck got me a place in the Whale . Sure I'm a good astronomer but so are lots of other guys. If I were ten years older, it would have been an honor, being picked for the first long jump in the first starship ever. At my age it was luck. "You'll want to know if the ship worked. Well, she did. Went like a bomb. We got lined up between Earth and Mars, you'll remember, and James pushed the button marked 'Jump'. Took his finger off the button and there we were: Alpha Centauri . Two months later your time, one second later by us. We covered our whole survey assignment like that, smooth as a pint of old and mild which right now I could certainly use. Better yet would be a pint of hot black coffee with sugar in. Failing that, I could even go for a long drink of cold water. There was never anything wrong with the Whale till right at the end and even then I doubt if it was the ship itself that fouled things up. "That was some survey assignment. We astronomers really lived. Wait till you see—but of course you won't. I could weep when I think of those miles of lovely color film, all gone up in smoke. "I'm shocked all right. I never said who I was. Matt Hennessy, from Farside Observatory, back of the Moon, just back from a proving flight cum astronomical survey in the starship Whale . Whoever you are who finds this tape, you're made. Take it to any radio station or newspaper office. You'll find you can name your price and don't take any wooden nickels. "Where had I got to? I'd told you how we happened to find Chang, hadn't I? That's what the natives called it. Walking, talking natives on a blue sky planet with 1.1 g gravity and a twenty per cent oxygen atmosphere at fifteen p.s.i. The odds against finding Chang on a six-sun survey on the first star jump ever must be up in the googols. We certainly were lucky. "The Chang natives aren't very technical—haven't got space travel for instance. They're good astronomers, though. We were able to show them our sun, in their telescopes. In their way, they're a highly civilized people. Look more like cats than people, but they're people all right. If you doubt it, chew these facts over. "One, they learned our language in four weeks. When I say they, I mean a ten-man team of them. "Two, they brew a near-beer that's a lot nearer than the canned stuff we had aboard the Whale . "Three, they've a great sense of humor. Ran rather to silly practical jokes, but still. Can't say I care for that hot-foot and belly-laugh stuff myself, but tastes differ. "Four, the ten-man language team also learned chess and table tennis. "But why go on? People who talk English, drink beer, like jokes and beat me at chess or table-tennis are people for my money, even if they look like tigers in trousers. "It was funny the way they won all the time at table tennis. They certainly weren't so hot at it. Maybe that ten per cent extra gravity put us off our strokes. As for chess, Svendlov was our champion. He won sometimes. The rest of us seemed to lose whichever Chingsi we played. There again it wasn't so much that they were good. How could they be, in the time? It was more that we all seemed to make silly mistakes when we played them and that's fatal in chess. Of course it's a screwy situation, playing chess with something that grows its own fur coat, has yellow eyes an inch and a half long and long white whiskers. Could you have kept your mind on the game? "And don't think I fell victim to their feline charm. The children were pets, but you didn't feel like patting the adults on their big grinning heads. Personally I didn't like the one I knew best. He was called—well, we called him Charley, and he was the ethnologist, ambassador, contact man, or whatever you like to call him, who came back with us. Why I disliked him was because he was always trying to get the edge on you. All the time he had to be top. Great sense of humor, of course. I nearly broke my neck on that butter-slide he fixed up in the metal alleyway to the Whale's engine room. Charley laughed fit to bust, everyone laughed, I even laughed myself though doing it hurt me more than the tumble had. Yes, life and soul of the party, old Charley ... "My last sight of the Minnow was a cabin full of dead and dying men, the sweetish stink of burned flesh and the choking reek of scorching insulation, the boat jolting and shuddering and beginning to break up, and in the middle of the flames, still unhurt, was Charley. He was laughing ... "My God, it's dark out here. Wonder how high I am. Must be all of fifty miles, and doing eight hundred miles an hour at least. I'll be doing more than that when I land. What's final velocity for a fifty-mile fall? Same as a fifty thousand mile fall, I suppose; same as escape; twenty-four thousand miles an hour. I'll make a mess ... "That's better. Why didn't I close my eyes before? Those star streaks made me dizzy. I'll make a nice shooting star when I hit air. Come to think of it, I must be deep in air now. Let's take a look. "It's getting lighter. Look at those peaks down there! Like great knives. I don't seem to be falling as fast as I expected though. Almost seem to be floating. Let's switch on the radio and tell the world hello. Hello, earth ... hello, again ... and good-by ... "Sorry about that. I passed out. I don't know what I said, if anything, and the suit recorder has no playback or eraser. What must have happened is that the suit ran out of oxygen, and I lost consciousness due to anoxia. I dreamed I switched on the radio, but I actually switched on the emergency tank, thank the Lord, and that brought me round. "Come to think of it, why not crack the suit and breath fresh air instead of bottled? "No. I'd have to get up to do that. I think I'll just lie here a little bit longer and get properly rested up before I try anything big like standing up. "I was telling about the return journey, wasn't I? The long jump back home, which should have dumped us between the orbits of Earth and Mars. Instead of which, when James took his finger off the button, the mass-detector showed nothing except the noise-level of the universe. "We were out in that no place for a day. We astronomers had to establish our exact position relative to the solar system. The crew had to find out exactly what went wrong. The physicists had to make mystic passes in front of meters and mutter about residual folds in stress-free space. Our task was easy, because we were about half a light-year from the sun. The crew's job was also easy: they found what went wrong in less than half an hour. "It still seems incredible. To program the ship for a star-jump, you merely told it where you were and where you wanted to go. In practical terms, that entailed first a series of exact measurements which had to be translated into the somewhat abstruse co-ordinate system we used based on the topological order of mass-points in the galaxy. Then you cut a tape on the computer and hit the button. Nothing was wrong with the computer. Nothing was wrong with the engines. We'd hit the right button and we'd gone to the place we'd aimed for. All we'd done was aim for the wrong place. It hurts me to tell you this and I'm just attached personnel with no space-flight tradition. In practical terms, one highly trained crew member had punched a wrong pattern of holes on the tape. Another equally skilled had failed to notice this when reading back. A childish error, highly improbable; twice repeated, thus squaring the improbability. Incredible, but that's what happened. "Anyway, we took good care with the next lot of measurements. That's why we were out there so long. They were cross-checked about five times. I got sick so I climbed into a spacesuit and went outside and took some photographs of the Sun which I hoped would help to determine hydrogen density in the outer regions. When I got back everything was ready. We disposed ourselves about the control room and relaxed for all we were worth. We were all praying that this time nothing would go wrong, and all looking forward to seeing Earth again after four months subjective time away, except for Charley, who was still chuckling and shaking his head, and Captain James who was glaring at Charley and obviously wishing human dignity permitted him to tear Charley limb from limb. Then James pressed the button. "Everything twanged like a bowstring. I felt myself turned inside out, passed through a small sieve, and poured back into shape. The entire bow wall-screen was full of Earth. Something was wrong all right, and this time it was much, much worse. We'd come out of the jump about two hundred miles above the Pacific, pointed straight down, traveling at a relative speed of about two thousand miles an hour. "It was a fantastic situation. Here was the Whale , the most powerful ship ever built, which could cover fifty light-years in a subjective time of one second, and it was helpless. For, as of course you know, the star-drive couldn't be used again for at least two hours. "The Whale also had ion rockets of course, the standard deuterium-fusion thing with direct conversion. As again you know, this is good for interplanetary flight because you can run it continuously and it has extremely high exhaust velocity. But in our situation it was no good because it has rather a low thrust. It would have taken more time than we had to deflect us enough to avoid a smash. We had five minutes to abandon ship. "James got us all into the Minnow at a dead run. There was no time to take anything at all except the clothes we stood in. The Minnow was meant for short heavy hops to planets or asteroids. In addition to the ion drive it had emergency atomic rockets, using steam for reaction mass. We thanked God for that when Cazamian canceled our downwards velocity with them in a few seconds. We curved away up over China and from about fifty miles high we saw the Whale hit the Pacific. Six hundred tons of mass at well over two thousand miles an hour make an almighty splash. By now you'll have divers down, but I doubt they'll salvage much you can use. "I wonder why James went down with the ship, as the saying is? Not that it made any difference. It must have broken his heart to know that his lovely ship was getting the chopper. Or did he suspect another human error? "We didn't have time to think about that, or even to get the radio working. The steam rockets blew up. Poor Cazamian was burnt to a crisp. Only thing that saved me was the spacesuit I was still wearing. I snapped the face plate down because the cabin was filling with fumes. I saw Charley coming out of the toilet—that's how he'd escaped—and I saw him beginning to laugh. Then the port side collapsed and I fell out. "I saw the launch spinning away, glowing red against a purplish black sky. I tumbled head over heels towards the huge curved shield of earth fifty miles below. I shut my eyes and that's about all I remember. I don't see how any of us could have survived. I think we're all dead. "I'll have to get up and crack this suit and let some air in. But I can't. I fell fifty miles without a parachute. I'm dead so I can't stand up." There was silence for a while except for the vicious howl of the wind. Then snow began to shift on the ledge. A man crawled stiffly out and came shakily to his feet. He moved slowly around for some time. After about two hours he returned to the hollow, squatted down and switched on the recorder. The voice began again, considerably wearier. "Hello there. I'm in the bleakest wilderness I've ever seen. This place makes the moon look cozy. There's precipice around me every way but one and that's up. So it's up I'll have to go till I find a way to go down. I've been chewing snow to quench my thirst but I could eat a horse. I picked up a short-wave broadcast on my suit but couldn't understand a word. Not English, not French, and there I stick. Listened to it for fifteen minutes just to hear a human voice again. I haven't much hope of reaching anyone with my five milliwatt suit transmitter but I'll keep trying. "Just before I start the climb there are two things I want to get on tape. The first is how I got here. I've remembered something from my military training, when I did some parachute jumps. Terminal velocity for a human body falling through air is about one hundred twenty m.p.h. Falling fifty miles is no worse than falling five hundred feet. You'd be lucky to live through a five hundred foot fall, true, but I've been lucky. The suit is bulky but light and probably slowed my fall. I hit a sixty mile an hour updraft this side of the mountain, skidded downhill through about half a mile of snow and fetched up in a drift. The suit is part worn but still operational. I'm fine. "The second thing I want to say is about the Chingsi, and here it is: watch out for them. Those jokers are dangerous. I'm not telling how because I've got a scientific reputation to watch. You'll have to figure it out for yourselves. Here are the clues: (1) The Chingsi talk and laugh but after all they aren't human. On an alien world a hundred light-years away, why shouldn't alien talents develop? A talent that's so uncertain and rudimentary here that most people don't believe it, might be highly developed out there. (2) The Whale expedition did fine till it found Chang. Then it hit a seam of bad luck. Real stinking bad luck that went on and on till it looks fishy. We lost the ship, we lost the launch, all but one of us lost our lives. We couldn't even win a game of ping-pong. "So what is luck, good or bad? Scientifically speaking, future chance events are by definition chance. They can turn out favorable or not. When a preponderance of chance events has occurred unfavorably, you've got bad luck. It's a fancy name for a lot of chance results that didn't go your way. But the gambler defines it differently. For him, luck refers to the future, and you've got bad luck when future chance events won't go your way. Scientific investigations into this have been inconclusive, but everyone knows that some people are lucky and others aren't. All we've got are hints and glimmers, the fumbling touch of a rudimentary talent. There's the evil eye legend and the Jonah, bad luck bringers. Superstition? Maybe; but ask the insurance companies about accident prones. What's in a name? Call a man unlucky and you're superstitious. Call him accident prone and that's sound business sense. I've said enough. "All the same, search the space-flight records, talk to the actuaries. When a ship is working perfectly and is operated by a hand-picked crew of highly trained men in perfect condition, how often is it wrecked by a series of silly errors happening one after another in defiance of probability? "I'll sign off with two thoughts, one depressing and one cheering. A single Chingsi wrecked our ship and our launch. What could a whole planetful of them do? "On the other hand, a talent that manipulates chance events is bound to be chancy. No matter how highly developed it can't be surefire. The proof is that I've survived to tell the tale." At twenty below zero and fifty miles an hour the wind ravaged the mountain. Peering through his polarized vizor at the white waste and the snow-filled air howling over it, sliding and stumbling with every step on a slope that got gradually steeper and seemed to go on forever, Matt Hennessy began to inch his way up the north face of Mount Everest. THE END Transcriber's Note: This etext was produced from Astounding Science Fiction February 1959. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed. Minor spelling and typographical errors have been corrected without note. | B. An astronomical surveyor who ended up there by accident |
Why did the Groacians hide the ship?
A. To overthrow the government.
B. They wanted to hide the Terrestrials as long as they could.
C. They were afraid to admit they knew where it was.
D. They wanted to keep it for further research.
| THE MADMAN FROM EARTH BY KEITH LAUMER You don't have to be crazy to be an earth diplomat—but on Groac it sure helps! [Transcriber's Note: This etext was produced from Worlds of If Science Fiction, March 1962. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed.] I "The Consul for the Terrestrial States," Retief said, "presents his compliments, et cetera, to the Ministry of Culture of the Groacian Autonomy, and with reference to the Ministry's invitation to attend a recital of interpretive grimacing, has the honor to express regret that he will be unable—" "You can't turn this invitation down," Administrative Assistant Meuhl said flatly. "I'll make that 'accepts with pleasure'." Retief exhaled a plume of cigar smoke. "Miss Meuhl," he said, "in the past couple of weeks I've sat through six light-concerts, four attempts at chamber music, and god knows how many assorted folk-art festivals. I've been tied up every off-duty hour since I got here—" "You can't offend the Groaci," Miss Meuhl said sharply. "Consul Whaffle would never have been so rude." "Whaffle left here three months ago," Retief said, "leaving me in charge." "Well," Miss Meuhl said, snapping off the dictyper. "I'm sure I don't know what excuse I can give the Minister." "Never mind the excuses," Retief said. "Just tell him I won't be there." He stood up. "Are you leaving the office?" Miss Meuhl adjusted her glasses. "I have some important letters here for your signature." "I don't recall dictating any letters today, Miss Meuhl," Retief said, pulling on a light cape. "I wrote them for you. They're just as Consul Whaffle would have wanted them." "Did you write all Whaffle's letters for him, Miss Meuhl?" "Consul Whaffle was an extremely busy man," Miss Meuhl said stiffly. "He had complete confidence in me." "Since I'm cutting out the culture from now on," Retief said, "I won't be so busy." "Well!" Miss Meuhl said. "May I ask where you'll be if something comes up?" "I'm going over to the Foreign Office Archives." Miss Meuhl blinked behind thick lenses. "Whatever for?" Retief looked thoughtfully at Miss Meuhl. "You've been here on Groac for four years, Miss Meuhl. What was behind the coup d'etat that put the present government in power?" "I'm sure I haven't pried into—" "What about that Terrestrial cruiser? The one that disappeared out this way about ten years back?" "Mr. Retief, those are just the sort of questions we avoid with the Groaci. I certainly hope you're not thinking of openly intruding—" "Why?" "The Groaci are a very sensitive race. They don't welcome outworlders raking up things. They've been gracious enough to let us live down the fact that Terrestrials subjected them to deep humiliation on one occasion." "You mean when they came looking for the cruiser?" "I, for one, am ashamed of the high-handed tactics that were employed, grilling these innocent people as though they were criminals. We try never to reopen that wound, Mr. Retief." "They never found the cruiser, did they?" "Certainly not on Groac." Retief nodded. "Thanks, Miss Meuhl," he said. "I'll be back before you close the office." Miss Meuhl's face was set in lines of grim disapproval as he closed the door. The pale-featured Groacian vibrated his throat-bladder in a distressed bleat. "Not to enter the Archives," he said in his faint voice. "The denial of permission. The deep regret of the Archivist." "The importance of my task here," Retief said, enunciating the glottal dialect with difficulty. "My interest in local history." "The impossibility of access to outworlders. To depart quietly." "The necessity that I enter." "The specific instructions of the Archivist." The Groacian's voice rose to a whisper. "To insist no longer. To give up this idea!" "OK, Skinny, I know when I'm licked," Retief said in Terran. "To keep your nose clean." Outside, Retief stood for a moment looking across at the deeply carved windowless stucco facades lining the street, then started off in the direction of the Terrestrial Consulate General. The few Groacians on the street eyed him furtively, veered to avoid him as he passed. Flimsy high-wheeled ground cars puffed silently along the resilient pavement. The air was clean and cool. At the office, Miss Meuhl would be waiting with another list of complaints. Retief studied the carving over the open doorways along the street. An elaborate one picked out in pinkish paint seemed to indicate the Groacian equivalent of a bar. Retief went in. A Groacian bartender was dispensing clay pots of alcoholic drink from the bar-pit at the center of the room. He looked at Retief and froze in mid-motion, a metal tube poised over a waiting pot. "To enjoy a cooling drink," Retief said in Groacian, squatting down at the edge of the pit. "To sample a true Groacian beverage." "To not enjoy my poor offerings," the Groacian mumbled. "A pain in the digestive sacs; to express regret." "To not worry," Retief said, irritated. "To pour it out and let me decide whether I like it." "To be grappled in by peace-keepers for poisoning of—foreigners." The barkeep looked around for support, found none. The Groaci customers, eyes elsewhere, were drifting away. "To get the lead out," Retief said, placing a thick gold-piece in the dish provided. "To shake a tentacle." "The procuring of a cage," a thin voice called from the sidelines. "The displaying of a freak." Retief turned. A tall Groacian vibrated his mandibles in a gesture of contempt. From his bluish throat coloration, it was apparent the creature was drunk. "To choke in your upper sac," the bartender hissed, extending his eyes toward the drunk. "To keep silent, litter-mate of drones." "To swallow your own poison, dispenser of vileness," the drunk whispered. "To find a proper cage for this zoo-piece." He wavered toward Retief. "To show this one in the streets, like all freaks." "Seen a lot of freaks like me, have you?" Retief asked, interestedly. "To speak intelligibly, malodorous outworlder," the drunk said. The barkeep whispered something, and two customers came up to the drunk, took his arms and helped him to the door. "To get a cage!" the drunk shrilled. "To keep the animals in their own stinking place." "I've changed my mind," Retief said to the bartender. "To be grateful as hell, but to have to hurry off now." He followed the drunk out the door. The other Groaci released him, hurried back inside. Retief looked at the weaving alien. "To begone, freak," the Groacian whispered. "To be pals," Retief said. "To be kind to dumb animals." "To have you hauled away to a stockyard, ill-odored foreign livestock." "To not be angry, fragrant native," Retief said. "To permit me to chum with you." "To flee before I take a cane to you!" "To have a drink together—" "To not endure such insolence!" The Groacian advanced toward Retief. Retief backed away. "To hold hands," Retief said. "To be palsy-walsy—" The Groacian reached for him, missed. A passer-by stepped around him, head down, scuttled away. Retief backed into the opening to a narrow crossway and offered further verbal familiarities to the drunken local, who followed, furious. Retief backed, rounded a corner into a narrow alley-like passage, deserted, silent ... except for the following Groacian. Retief stepped around him, seized his collar and yanked. The Groacian fell on his back. Retief stood over him. The downed native half-rose; Retief put a foot against his chest and pushed. "To not be going anywhere for a few minutes," Retief said. "To stay right here and have a nice long talk." II "There you are!" Miss Meuhl said, eyeing Retief over her lenses. "There are two gentlemen waiting to see you. Groacian gentlemen." "Government men, I imagine. Word travels fast." Retief pulled off his cape. "This saves me the trouble of paying another call at the Foreign Ministry." "What have you been doing? They seem very upset, I don't mind telling you." "I'm sure you don't. Come along. And bring an official recorder." Two Groaci wearing heavy eye-shields and elaborate crest ornaments indicative of rank rose as Retief entered the room. Neither offered a courteous snap of the mandibles, Retief noted. They were mad, all right. "I am Fith, of the Terrestrial Desk, Ministry of Foreign Affairs, Mr. Consul," the taller Groacian said, in lisping Terran. "May I present Shluh, of the Internal Police?" "Sit down, gentlemen," Retief said. They resumed their seats. Miss Meuhl hovered nervously, then sat on the edge of a comfortless chair. "Oh, it's such a pleasure—" she began. "Never mind that," Retief said. "These gentlemen didn't come here to sip tea today." "So true," Fith said. "Frankly, I have had a most disturbing report, Mr. Consul. I shall ask Shluh to recount it." He nodded to the police chief. "One hour ago," The Groacian said, "a Groacian national was brought to hospital suffering from serious contusions. Questioning of this individual revealed that he had been set upon and beaten by a foreigner. A Terrestrial, to be precise. Investigation by my department indicates that the description of the culprit closely matches that of the Terrestrial Consul." Miss Meuhl gasped audibly. "Have you ever heard," Retief said, looking steadily at Fith, "of a Terrestrial cruiser, the ISV Terrific , which dropped from sight in this sector nine years ago?" "Really!" Miss Meuhl exclaimed, rising. "I wash my hands—" "Just keep that recorder going," Retief snapped. "I'll not be a party—" "You'll do as you're told, Miss Meuhl," Retief said quietly. "I'm telling you to make an official sealed record of this conversation." Miss Meuhl sat down. Fith puffed out his throat indignantly. "You reopen an old wound, Mr. Consul. It reminds us of certain illegal treatment at Terrestrial hands—" "Hogwash," Retief said. "That tune went over with my predecessors, but it hits a sour note with me." "All our efforts," Miss Meuhl said, "to live down that terrible episode! And you—" "Terrible? I understand that a Terrestrial task force stood off Groac and sent a delegation down to ask questions. They got some funny answers, and stayed on to dig around a little. After a week they left. Somewhat annoying to the Groaci, maybe—at the most. If they were innocent." "IF!" Miss Meuhl burst out. "If, indeed!" Fith said, his weak voice trembling. "I must protest your—" "Save the protests, Fith. You have some explaining to do. And I don't think your story will be good enough." "It is for you to explain! This person who was beaten—" "Not beaten. Just rapped a few times to loosen his memory." "Then you admit—" "It worked, too. He remembered lots of things, once he put his mind to it." Fith rose; Shluh followed suit. "I shall ask for your immediate recall, Mr. Consul. Were it not for your diplomatic immunity, I should do more—" "Why did the government fall, Fith? It was just after the task force paid its visit, and before the arrival of the first Terrestrial diplomatic mission." "This is an internal matter!" Fith cried, in his faint Groacian voice. "The new regime has shown itself most amiable to you Terrestrials. It has outdone itself—" "—to keep the Terrestrial consul and his staff in the dark," Retief said. "And the same goes for the few terrestrial businessmen you've visaed. This continual round of culture; no social contacts outside the diplomatic circle; no travel permits to visit out-lying districts, or your satellite—" "Enough!" Fith's mandibles quivered in distress. "I can talk no more of this matter—" "You'll talk to me, or there'll be a task force here in five days to do the talking," Retief said. "You can't!" Miss Meuhl gasped. Retief turned a steady look on Miss Meuhl. She closed her mouth. The Groaci sat down. "Answer me this one," Retief said, looking at Shluh. "A few years back—about nine, I think—there was a little parade held here. Some curious looking creatures were captured. After being securely caged, they were exhibited to the gentle Groaci public. Hauled through the streets. Very educational, no doubt. A highly cultural show. "Funny thing about these animals. They wore clothes. They seemed to communicate with each other. Altogether it was a very amusing exhibit. "Tell me, Shluh, what happened to those six Terrestrials after the parade was over?" Fith made a choked noise and spoke rapidly to Shluh in Groacian. Shluh retracted his eyes, shrank down in his chair. Miss Meuhl opened her mouth, closed it and blinked rapidly. "How did they die?" Retief snapped. "Did you murder them, cut their throats, shoot them or bury them alive? What amusing end did you figure out for them? Research, maybe? Cut them open to see what made them yell...." "No!" Fith gasped. "I must correct this terrible false impression at once." "False impression, hell," Retief said. "They were Terrans! A simple narco-interrogation would get that out of any Groacian who saw the parade." "Yes," Fith said weakly. "It is true, they were Terrestrials. But there was no killing." "They're alive?" "Alas, no. They ... died." Miss Meuhl yelped faintly. "I see," Retief said. "They died." "We tried to keep them alive, of course. But we did not know what foods—" "Didn't take the trouble to find out, either, did you?" "They fell ill," Fith said. "One by one...." "We'll deal with that question later," Retief said. "Right now, I want more information. Where did you get them? Where did you hide the ship? What happened to the rest of the crew? Did they 'fall ill' before the big parade?" "There were no more! Absolutely, I assure you!" "Killed in the crash landing?" "No crash landing. The ship descended intact, east of the city. The ... Terrestrials ... were unharmed. Naturally, we feared them. They were strange to us. We had never before seen such beings." "Stepped off the ship with guns blazing, did they?" "Guns? No, no guns—" "They raised their hands, didn't they? Asked for help. You helped them; helped them to death." "How could we know?" Fith moaned. "How could you know a flotilla would show up in a few months looking for them, you mean? That was a shock, wasn't it? I'll bet you had a brisk time of it hiding the ship, and shutting everybody up. A close call, eh?" "We were afraid," Shluh said. "We are a simple people. We feared the strange creatures from the alien craft. We did not kill them, but we felt it was as well they ... did not survive. Then, when the warships came, we realized our error. But we feared to speak. We purged our guilty leaders, concealed what had happened, and ... offered our friendship. We invited the opening of diplomatic relations. We made a blunder, it is true, a great blunder. But we have tried to make amends...." "Where is the ship?" "The ship?" "What did you do with it? It was too big to just walk off and forget. Where is it?" The two Groacians exchanged looks. "We wish to show our contrition," Fith said. "We will show you the ship." "Miss Meuhl," Retief said. "If I don't come back in a reasonable length of time, transmit that recording to Regional Headquarters, sealed." He stood, looked at the Groaci. "Let's go," he said. Retief stooped under the heavy timbers shoring the entry to the cavern. He peered into the gloom at the curving flank of the space-burned hull. "Any lights in here?" he asked. A Groacian threw a switch. A weak bluish glow sprang up. Retief walked along the raised wooden catwalk, studying the ship. Empty emplacements gaped below lensless scanner eyes. Littered decking was visible within the half-open entry port. Near the bow the words 'IVS Terrific B7 New Terra' were lettered in bright chrome duralloy. "How did you get it in here?" Retief asked. "It was hauled here from the landing point, some nine miles distant," Fith said, his voice thinner than ever. "This is a natural crevasse. The vessel was lowered into it and roofed over." "How did you shield it so the detectors didn't pick it up?" "All here is high-grade iron ore," Fith said, waving a member. "Great veins of almost pure metal." Retief grunted. "Let's go inside." Shluh came forward with a hand-lamp. The party entered the ship. Retief clambered up a narrow companionway, glanced around the interior of the control compartment. Dust was thick on the deck, the stanchions where acceleration couches had been mounted, the empty instrument panels, the litter of sheared bolts, scraps of wire and paper. A thin frosting of rust dulled the exposed metal where cutting torches had sliced away heavy shielding. There was a faint odor of stale bedding. "The cargo compartment—" Shluh began. "I've seen enough," Retief said. Silently, the Groacians led the way back out through the tunnel and into the late afternoon sunshine. As they climbed the slope to the steam car, Fith came to Retief's side. "Indeed, I hope that this will be the end of this unfortunate affair," he said. "Now that all has been fully and honestly shown—" "You can skip all that," Retief said. "You're nine years late. The crew was still alive when the task force called, I imagine. You killed them—or let them die—rather than take the chance of admitting what you'd done." "We were at fault," Fith said abjectly. "Now we wish only friendship." "The Terrific was a heavy cruiser, about twenty thousand tons." Retief looked grimly at the slender Foreign Office official. "Where is she, Fith? I won't settle for a hundred-ton lifeboat." Fith erected his eye stalks so violently that one eye-shield fell off. "I know nothing of ... of...." He stopped. His throat vibrated rapidly as he struggled for calm. "My government can entertain no further accusations, Mr. Consul," he said at last. "I have been completely candid with you, I have overlooked your probing into matters not properly within your sphere of responsibility. My patience is at an end." "Where is that ship?" Retief rapped out. "You never learn, do you? You're still convinced you can hide the whole thing and forget it. I'm telling you you can't." "We return to the city now," Fith said. "I can do no more." "You can and you will, Fith," Retief said. "I intend to get to the truth of this matter." Fith spoke to Shluh in rapid Groacian. The police chief gestured to his four armed constables. They moved to ring Retief in. Retief eyed Fith. "Don't try it," he said. "You'll just get yourself in deeper." Fith clacked his mandibles angrily, eye stalks canted aggressively toward the Terrestrial. "Out of deference to your diplomatic status, Terrestrial, I shall ignore your insulting remarks," Fith said in his reedy voice. "Let us now return to the city." Retief looked at the four policemen. "I see your point," he said. Fith followed him into the car, sat rigidly at the far end of the seat. "I advise you to remain very close to your consulate," Fith said. "I advise you to dismiss these fancies from your mind, and to enjoy the cultural aspects of life at Groac. Especially, I should not venture out of the city, or appear overly curious about matters of concern only to the Groacian government." In the front seat, Shluh looked straight ahead. The loosely-sprung vehicle bobbed and swayed along the narrow highway. Retief listened to the rhythmic puffing of the motor and said nothing. III "Miss Meuhl," Retief said, "I want you to listen carefully to what I'm going to tell you. I have to move rapidly now, to catch the Groaci off guard." "I'm sure I don't know what you're talking about," Miss Meuhl snapped, her eyes sharp behind the heavy lenses. "If you'll listen, you may find out," Retief said. "I have no time to waste, Miss Meuhl. They won't be expecting an immediate move—I hope—and that may give me the latitude I need." "You're still determined to make an issue of that incident!" Miss Meuhl snorted. "I really can hardly blame the Groaci. They are not a sophisticated race; they had never before met aliens." "You're ready to forgive a great deal, Miss Meuhl. But it's not what happened nine years ago I'm concerned with. It's what's happening now. I've told you that it was only a lifeboat the Groaci have hidden out. Don't you understand the implication? That vessel couldn't have come far. The cruiser itself must be somewhere near by. I want to know where!" "The Groaci don't know. They're a very cultured, gentle people. You can do irreparable harm to the reputation of Terrestrials if you insist—" "That's my decision," Retief said. "I have a job to do and we're wasting time." He crossed the room to his desk, opened a drawer and took out a slim-barreled needler. "This office is being watched. Not very efficiently, if I know the Groaci. I think I can get past them all right." "Where are you going with ... that?" Miss Meuhl stared at the needler. "What in the world—" "The Groaci won't waste any time destroying every piece of paper in their files relating to this thing. I have to get what I need before it's too late. If I wait for an official Inquiry Commission, they'll find nothing but blank smiles." "You're out of your mind!" Miss Meuhl stood up, quivering with indignation. "You're like a ... a...." "You and I are in a tight spot, Miss Meuhl. The logical next move for the Groaci is to dispose of both of us. We're the only ones who know what happened. Fith almost did the job this afternoon, but I bluffed him out—for the moment." Miss Meuhl emitted a shrill laugh. "Your fantasies are getting the better of you," she gasped. "In danger, indeed! Disposing of me! I've never heard anything so ridiculous." "Stay in this office. Close and safe-lock the door. You've got food and water in the dispenser. I suggest you stock up, before they shut the supply down. Don't let anyone in, on any pretext whatever. I'll keep in touch with you via hand-phone." "What are you planning to do?" "If I don't make it back here, transmit the sealed record of this afternoon's conversation, along with the information I've given you. Beam it through on a mayday priority. Then tell the Groaci what you've done and sit tight. I think you'll be all right. It won't be easy to blast in here and anyway, they won't make things worse by killing you. A force can be here in a week." "I'll do nothing of the sort! The Groaci are very fond of me! You ... Johnny-come-lately! Roughneck! Setting out to destroy—" "Blame it on me if it will make you feel any better," Retief said, "but don't be fool enough to trust them." He pulled on a cape, opened the door. "I'll be back in a couple of hours," he said. Miss Meuhl stared after him silently as he closed the door. It was an hour before dawn when Retief keyed the combination to the safe-lock and stepped into the darkened consular office. He looked tired. Miss Meuhl, dozing in a chair, awoke with a start. She looked at Retief, rose and snapped on a light, turned to stare. "What in the world—Where have you been? What's happened to your clothing?" "I got a little dirty. Don't worry about it." Retief went to his desk, opened a drawer and replaced the needler. "Where have you been?" Miss Meuhl demanded. "I stayed here—" "I'm glad you did," Retief said. "I hope you piled up a supply of food and water from the dispenser, too. We'll be holed up here for a week, at least." He jotted figures on a pad. "Warm up the official sender. I have a long transmission for Regional Headquarters." "Are you going to tell me where you've been?" "I have a message to get off first, Miss Meuhl," Retief said sharply. "I've been to the Foreign Ministry," he added. "I'll tell you all about it later." "At this hour? There's no one there...." "Exactly." Miss Meuhl gasped. "You mean you broke in? You burgled the Foreign Office?" "That's right," Retief said calmly. "Now—" "This is absolutely the end!" Miss Meuhl said. "Thank heaven I've already—" "Get that sender going, woman!" Retief snapped. "This is important." "I've already done so, Mr. Retief!" Miss Meuhl said harshly. "I've been waiting for you to come back here...." She turned to the communicator, flipped levers. The screen snapped aglow, and a wavering long-distance image appeared. "He's here now," Miss Meuhl said to the screen. She looked at Retief triumphantly. "That's good," Retief said. "I don't think the Groaci can knock us off the air, but—" "I have done my duty, Mr. Retief," Miss Meuhl said. "I made a full report to Regional Headquarters last night, as soon as you left this office. Any doubts I may have had as to the rightness of that decision have been completely dispelled by what you've just told me." Retief looked at her levelly. "You've been a busy girl, Miss Meuhl. Did you mention the six Terrestrials who were killed here?" "That had no bearing on the matter of your wild behavior! I must say, in all my years in the Corps, I've never encountered a personality less suited to diplomatic work." The screen crackled, the ten-second transmission lag having elapsed. "Mr. Retief," the face on the screen said, "I am Counsellor Pardy, DSO-1, Deputy Under-secretary for the region. I have received a report on your conduct which makes it mandatory for me to relieve you administratively, vice Miss Yolanda Meuhl, DAO-9. Pending the findings of a Board of Inquiry, you will—" Retief reached out and snapped off the communicator. The triumphant look faded from Miss Meuhl's face. "Why, what is the meaning—" "If I'd listened any longer, I might have heard something I couldn't ignore. I can't afford that, at this moment. Listen, Miss Meuhl," Retief went on earnestly, "I've found the missing cruiser." "You heard him relieve you!" "I heard him say he was going to, Miss Meuhl. But until I've heard and acknowledged a verbal order, it has no force. If I'm wrong, he'll get my resignation. If I'm right, that suspension would be embarrassing all around." "You're defying lawful authority! I'm in charge here now." Miss Meuhl stepped to the local communicator. "I'm going to report this terrible thing to the Groaci at once, and offer my profound—" "Don't touch that screen," Retief said. "You go sit in that corner where I can keep an eye on you. I'm going to make a sealed tape for transmission to Headquarters, along with a call for an armed task force. Then we'll settle down to wait." Retief ignored Miss Meuhl's fury as he spoke into the recorder. The local communicator chimed. Miss Meuhl jumped up, staring at it. "Go ahead," Retief said. "Answer it." A Groacian official appeared on the screen. "Yolanda Meuhl," he said without preamble, "for the Foreign Minister of the Groacian Autonomy, I herewith accredit you as Terrestrial Consul to Groac, in accordance with the advices transmitted to my government direct from the Terrestrial Headquarters. As consul, you are requested to make available for questioning Mr. J. Retief, former consul, in connection with the assault on two peace keepers and illegal entry into the offices of the Ministry for Foreign Affairs." "Why, why," Miss Meuhl stammered. "Yes, of course. And I do want to express my deepest regrets—" Retief rose, went to the communicator, assisted Miss Meuhl aside. "Listen carefully, Fith," he said. "Your bluff has been called. You don't come in and we don't come out. Your camouflage worked for nine years, but it's all over now. I suggest you keep your heads and resist the temptation to make matters worse than they are." "Miss Meuhl," Fith said, "a peace squad waits outside your consulate. It is clear you are in the hands of a dangerous lunatic. As always, the Groaci wish only friendship with the Terrestrials, but—" "Don't bother," Retief said. "You know what was in those files I looked over this morning." Retief turned at a sound behind him. Miss Meuhl was at the door, reaching for the safe-lock release.... "Don't!" Retief jumped—too late. The door burst inward. A crowd of crested Groaci pressed into the room, pushed Miss Meuhl back, aimed scatter guns at Retief. Police Chief Shluh pushed forward. "Attempt no violence, Terrestrial," he said. "I cannot promise to restrain my men." "You're violating Terrestrial territory, Shluh," Retief said steadily. "I suggest you move back out the same way you came in." "I invited them here," Miss Meuhl spoke up. "They are here at my express wish." "Are they? Are you sure you meant to go this far, Miss Meuhl? A squad of armed Groaci in the consulate?" "You are the consul, Miss Yolanda Meuhl," Shluh said. "Would it not be best if we removed this deranged person to a place of safety?" "You're making a serious mistake, Shluh," Retief said. "Yes," Miss Meuhl said. "You're quite right, Mr. Shluh. Please escort Mr. Retief to his quarters in this building—" "I don't advise you to violate my diplomatic immunity, Fith," Retief said. "As chief of mission," Miss Meuhl said quickly, "I hereby waive immunity in the case of Mr. Retief." Shluh produced a hand recorder. "Kindly repeat your statement, Madam, officially," he said. "I wish no question to arise later." "Don't be a fool, woman," Retief said. "Don't you see what you're letting yourself in for? This would be a hell of a good time for you to figure out whose side you're on." "I'm on the side of common decency!" "You've been taken in. These people are concealing—" "You think all women are fools, don't you, Mr. Retief?" She turned to the police chief and spoke into the microphone he held up. "That's an illegal waiver," Retief said. "I'm consul here, whatever rumors you've heard. This thing's coming out into the open, whatever you do. Don't add violation of the Consulate to the list of Groacian atrocities." "Take the man," Shluh said. | C. They were afraid to admit they knew where it was. |
Why was Macklin's wife hysterical when she called to speak with Ferris and Mitchell?
A. Her husband was very ill from the virus
B. Her husband was still having headaches
C. She thought they had given her husband heroin.
D. Her husband's blood pressure had dropped extremely low.
| THE BIG HEADACHE BY JIM HARMON What's the principal cause of headaches? Why, having a head, of course! [Transcriber's Note: This etext was produced from Worlds of If Science Fiction, September 1962. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed.] I "Do you think we'll have to use force on Macklin to get him to cooperate in the experiment?" Ferris asked eagerly. "How are you going to go about forcing him, Doctor?" Mitchell inquired. "He outweighs you by fifty pounds and you needn't look to me for help against that repatriated fullback." Ferris fingered the collar of his starched lab smock. "Guess I got carried away for a moment. But Macklin is exactly what we need for a quick, dramatic test. We've had it if he turns us down." "I know," Mitchell said, exhaling deeply. "Somehow the men with the money just can't seem to understand basic research. Who would have financed a study of cyclic periods of the hedgehog? Yet the information gained from that study is vital in cancer research." "When we prove our results that should be of enough practical value for anyone. But those crummy trustees didn't even leave us enough for a field test." Ferris scrubbed his thin hand over the bony ridge of his forehead. "I've been worrying so much about this I've got the ancestor of all headaches." Mitchell's blue eyes narrowed and his boyish face took on an expression of demonic intensity. "Ferris, would you consider—?" "No!" the smaller man yelled. "You can't expect me to violate professional ethics and test my own discovery on myself." " Our discovery," Mitchell said politely. "That's what I meant to say. But I'm not sure it would be completely ethical with even a discovery partly mine." "You're right. Besides who cares if you or I are cured of headaches? Our reputations don't go outside our own fields," Mitchell said. "But now Macklin—" Elliot Macklin had inherited the reputation of the late Albert Einstein in the popular mind. He was the man people thought of when the word "mathematician" or even "scientist" was mentioned. No one knew whether his Theory of Spatium was correct or not because no one had yet been able to frame an argument with it. Macklin was in his early fifties but looked in his late thirties, with the build of a football player. The government took up a lot of his time using him as the symbol of the Ideal Scientist to help recruit Science and Engineering Cadets. For the past seven years Macklin—who was the Advanced Studies Department of Firestone University—had been involved in devising a faster-than-light drive to help the Army reach Pluto and eventually the nearer stars. Mitchell had overheard two coeds talking and so knew that the project was nearing completion. If so, it was a case of Ad astra per aspirin . The only thing that could delay the project was Macklin's health. Despite his impressive body, some years before he had suffered a mild stroke ... or at least a vascular spasm of a cerebral artery. It was known that he suffered from the vilest variety of migraine. A cycle of the headaches had caused him to be absent from his classes for several weeks, and there were an unusual number of military uniforms seen around the campus. Ferris paced off the tidy measurements of the office outside the laboratory in the biology building. Mitchell sat slumped in the chair behind the blond imitation wood desk, watching him disinterestedly. "Do you suppose the Great Man will actually show up?" Ferris demanded, pausing in mid-stride. "I imagine he will," Mitchell said. "Macklin's always seemed a decent enough fellow when I've had lunch with him or seen him at the trustees meetings." "He's always treated me like dirt," Ferris said heatedly. "Everyone on this campus treats biologists like dirt. Sometimes I want to bash in their smug faces." Sometimes, Mitchell reflected, Ferris displayed a certain lack of scientific detachment. There came a discreet knock on the door. "Please come in," Mitchell said. Elliot Macklin entered in a cloud of pipe smoke and a tweed jacket. He looked more than a little like a postgraduate student, and Mitchell suspected that that was his intention. He shook hands warmly with Mitchell. "Good of you to ask me over, Steven." Macklin threw a big arm across Ferris' shoulders. "How have you been, Harold?" Ferris' face flickered between pink and white. "Fine, thank you, doctor." Macklin dropped on the edge of the desk and adjusted his pipe. "Now what's this about you wanting my help on something? And please keep the explanation simple. Biology isn't my field, you know." Mitchell moved around the desk casually. "Actually, Doctor, we haven't the right to ask this of a man of your importance. There may be an element of risk." The mathematician clamped onto his pipe and showed his teeth. "Now you have me intrigued. What is it all about?" "Doctor, we understand you have severe headaches," Mitchell said. Macklin nodded. "That's right, Steven. Migraine." "That must be terrible," Ferris said. "All your fine reputation and lavish salary can't be much consolation when that ripping, tearing agony begins, can it?" "No, Harold, it isn't," Macklin admitted. "What does your project have to do with my headaches?" "Doctor," Mitchell said, "what would you say the most common complaint of man is?" "I would have said the common cold," Macklin replied, "but I suppose from what you have said you mean headaches." "Headaches," Mitchell agreed. "Everybody has them at some time in his life. Some people have them every day. Some are driven to suicide by their headaches." "Yes," Macklin said. "But think," Ferris interjected, "what a boon it would be if everyone could be cured of headaches forever by one simple injection." "I don't suppose the manufacturers of aspirin would like you. But it would please about everybody else." "Aspirins would still be used to reduce fever and relieve muscular pains," Mitchell said. "I see. Are you two saying you have such a shot? Can you cure headaches?" "We think we can," Ferris said. "How can you have a specific for a number of different causes?" Macklin asked. "I know that much about the subject." "There are a number of different causes for headaches—nervous strain, fatigue, physical diseases from kidney complaints to tumors, over-indulgence—but there is one effect of all of this, the one real cause of headaches," Mitchell announced. "We have definitely established this for this first time," Ferris added. "That's fine," Macklin said, sucking on his pipe. "And this effect that produces headaches is?" "The pressure effect caused by pituitrin in the brain," Mitchell said eagerly. "That is, the constriction of blood vessels in the telencephalon section of the frontal lobes. It's caused by an over-production of the pituitary gland. We have artificially bred a virus that feeds on pituitrin." "That may mean the end of headaches, but I would think it would mean the end of the race as well," Macklin said. "In certain areas it is valuable to have a constriction of blood vessels." "The virus," Ferris explained, "can easily be localized and stabilized. A colony of virus in the brain cells will relax the cerebral vessels—and only the cerebral vessels—so that the cerebrospinal fluid doesn't create pressure in the cavities of the brain." The mathematician took the pipe out of his mouth. "If this really works, I could stop using that damned gynergen, couldn't I? The stuff makes me violently sick to my stomach. But it's better than the migraine. How should I go about removing my curse?" He reinserted the pipe. "I assure you, you can forget ergotamine tartrate," Ferris said. "Our discovery will work." "Will work," Macklin said thoughtfully. "The operative word. It hasn't worked then?" "Certainly it has," Ferris said. "On rats, on chimps...." "But not on humans?" Macklin asked. "Not yet," Mitchell admitted. "Well," Macklin said. "Well." He thumped pipe ashes out into his palm. "Certainly you can get volunteers. Convicts. Conscientious objectors from the Army." "We want you," Ferris told him. Macklin coughed. "I don't want to overestimate my value but the government wouldn't like it very well if I died in the middle of this project. My wife would like it even less." Ferris turned his back on the mathematician. Mitchell could see him mouthing the word yellow . "Doctor," Mitchell said quickly, "I know it's a tremendous favor to ask of a man of your position. But you can understand our problem. Unless we can produce quick, conclusive and dramatic proof of our studies we can get no more financial backing. We should run a large-scale field test. But we haven't the time or money for that. We can cure the headaches of one person and that's the limit of our resources." "I'm tempted," Macklin said hesitantly, "but the answer is go. I mean ' no '. I'd like to help you out, but I'm afraid I owe too much to others to take the rest—the risk, I mean." Macklin ran the back of his knuckles across his forehead. "I really would like to take you up on it. When I start making slips like that it means another attack of migraine. The drilling, grinding pain through my temples and around my eyeballs. The flashes of light, the rioting pools of color playing on the back of my lids. Ugh." Ferris smiled. "Gynergen makes you sick, does it, doctor? Produces nausea, eh? The pain of that turns you almost wrong side out, doesn't it? You aren't much better off with it than without, are you? I've heard some say they preferred the migraine." Macklin carefully arranged his pipe along with the tools he used to tend it in a worn leather case. "Tell me," he said, "what is the worst that could happen to me?" "Low blood pressure," Ferris said. "That's not so bad," Macklin said. "How low can it get?" "When your heart stops, your blood pressure goes to its lowest point," Mitchell said. A dew of perspiration had bloomed on Macklin's forehead. "Is there much risk of that?" "Practically none," Mitchell said. "We have to give you the worst possibilities. All our test animals survived and seem perfectly happy and contented. As I said, the virus is self-stabilizing. Ferris and I are confident that there is no danger.... But we may be wrong." Macklin held his head in both hands. "Why did you two select me ?" "You're an important man, doctor," Ferris said. "Nobody would care if Mitchell or I cured ourselves of headaches—they might not even believe us if we said we did. But the proper authorities will believe a man of your reputation. Besides, neither of us has a record of chronic migraine. You do." "Yes, I do," Macklin said. "Very well. Go ahead. Give me your injection." Mitchell cleared his throat. "Are you positive, doctor?" he asked uncertainly. "Perhaps you would like a few days to think it over." "No! I'm ready. Go ahead, right now." "There's a simple release," Ferris said smoothly. Macklin groped in his pocket for a pen. II "Ferris!" Mitchell yelled, slamming the laboratory door behind him. "Right here," the small man said briskly. He was sitting at a work table, penciling notes. "I've been expecting you." "Doctor—Harold—you shouldn't have given this story to the newspapers," Mitchell said. He tapped the back of his hand against the folded paper. "On the contrary, I should and I did," Ferris answered. "We wanted something dramatic to show to the trustees and here it is." "Yes, we wanted to show our proof to the trustees—but not broadcast unverified results to the press. It's too early for that!" "Don't be so stuffy and conservative, Mitchell! Macklin's cured, isn't he? By established periodic cycle he should be suffering hell right now, shouldn't he? But thanks to our treatment he is perfectly happy, with no unfortunate side effects such as gynergen produces." "It's a significant test case, yes. But not enough to go to the newspapers with. If it wasn't enough to go to the press with, it wasn't enough to try and breach the trustees with. Don't you see? The public will hand down a ukase demanding our virus, just as they demanded the Salk vaccine and the Grennell serum." "But—" The shrill call of the telephone interrupted Mitchell's objections. Ferris excused himself and crossed to the instrument. He answered it and listened for a moment, his face growing impatient. "It's Macklin's wife," Ferris said. "Do you want to talk to her? I'm no good with hysterical women." "Hysterical?" Mitchell muttered in alarm and went to the phone. "Hello?" Mitchell said reluctantly. "Mrs. Macklin?" "You are the other one," the clear feminine voice said. "Your name is Mitchell." She couldn't have sounded calmer or more self-possessed, Mitchell thought. "That's right, Mrs. Macklin. I'm Dr. Steven Mitchell, Dr. Ferris's associate." "Do you have a license to dispense narcotics?" "What do you mean by that, Mrs. Macklin," Mitchell said sharply. "I used to be a nurse, Dr. Mitchell. I know you've given my husband heroin." "That's absurd. What makes you think a thing like that?" "The—trance he's in now." "Now, Mrs. Macklin. Neither Dr. Ferris or myself have been near your husband for a full day. The effects of a narcotic would have worn off by this time." "Most known narcotics," she admitted, "but evidently you have discovered something new. Is it so expensive to refine you and Ferris have to recruit new customers to keep yourselves supplied?" "Mrs. Macklin! I think I had better talk to you later when you are calmer." Mitchell dropped the receiver heavily. "What could be wrong with Macklin?" he asked without removing his hand from the telephone. Ferris frowned, making quotation marks above his nose. "Let's have a look at the test animals." Together they marched over to the cages and peered through the honeycomb pattern of the wire. The test chimp, Dean, was sitting peacefully in a corner scratching under his arms with the back of his knuckles. Jerry, their control in the experiment, who was practically Dean's twin except that he had received no injection of the E-M Virus, was stomping up and down punching his fingers through the wire, worrying the lock on the cage. "Jerry is a great deal more active than Dean," Mitchell said. "Yes, but Dean isn't sick. He just doesn't seem to have as much nervous energy to burn up. Nothing wrong with his thyroid either." They went to the smaller cages. They found the situation with the rats, Bud and Lou, much the same. "I don't know. Maybe they just have tired blood," Mitchell ventured. "Iron deficiency anemia?" "Never mind, doctor. It was a form of humor. I think we had better see exactly what is wrong with Elliot Macklin." "There's nothing wrong with him," Ferris snapped. "He's probably just trying to get us in trouble, the ingrate!" Macklin's traditional ranch house was small but attractive in aqua-tinted aluminum. Under Mitchell's thumb the bell chimbed dum-de-de-dum-dum-dum . As they waited Mitchell glanced at Ferris. He seemed completely undisturbed, perhaps slightly curious. The door unlatched and swung back. "Mrs. Macklin," Mitchell said quickly, "I'm sure we can help if there is anything wrong with your husband. This is Dr. Ferris. I am Dr. Mitchell." "You had certainly better help him, gentlemen." She stood out of the doorway for them to pass. Mrs. Macklin was an attractive brunette in her late thirties. She wore an expensive yellow dress. And she had a sharp-cornered jawline. The Army officer came out into the hall to meet them. "You are the gentlemen who gave Dr. Macklin the unauthorized injection," he said. It wasn't a question. "I don't like that 'unauthorized'," Ferris snapped. The colonel—Mitchell spotted the eagles on his green tunic—lifted a heavy eyebrow. "No? Are you medical doctors? Are you authorized to treat illnesses?" "We weren't treating an illness," Mitchell said. "We were discovering a method of treatment. What concern is it of yours?" The colonel smiled thinly. "Dr. Macklin is my concern. And everything that happens to him. The Army doesn't like what you have done to him." Mitchell wondered desperately just what they had done to the man. "Can we see him?" Mitchell asked. "Why not? You can't do much worse than murder him now. That might be just as well. We have laws to cover that." The colonel led them into the comfortable, over-feminine living room. Macklin sat in an easy chair draped in embroidery, smoking. Mitchell suddenly realized Macklin used a pipe as a form of masculine protest to his home surroundings. On the coffee table in front of Macklin were some odd-shaped building blocks such as were used in nursery schools. A second uniformed man—another colonel but with the snake-entwined staff of the medical corps in his insignia—was kneeling at the table on the marble-effect carpet. The Army physician stood up and brushed his knees, undusted from the scrupulously clean rug. "What's wrong with him, Sidney?" the other officer asked the doctor. "Not a thing," Sidney said. "He's the healthiest, happiest, most well-adjusted man I've ever examined, Carson." "But—" Colonel Carson protested. "Oh, he's changed all right," the Army doctor answered. "He's not the same man as he used to be." "How is he different?" Mitchell demanded. The medic examined Mitchell and Ferris critically before answering. "He used to be a mathematical genius." "And now?" Mitchell said impatiently. "Now he is a moron," the medic said. III Mitchell tried to stop Colonel Sidney as he went past, but the doctor mumbled he had a report to make. Mitchell and Ferris stared at Colonel Carson and Macklin and at each other. "What did he mean, Macklin is an idiot?" Mitchell asked. "Not an idiot," Colonel Carson corrected primly. "Dr. Macklin is a moron. He's legally responsible, but he's extremely stupid." "I'm not so dumb," Macklin said defensively. "I beg your pardon, sir," Carson said. "I didn't intend any offense. But according to all the standard intelligence tests we have given you, your clinical intelligence quotient is that of a moron." "That's just on book learning," Macklin said. "There's a lot you learn in life that you don't get out of books, son." "I'm confident that's true, sir," Colonel Carson said. He turned to the two biologists. "Perhaps we had better speak outside." "But—" Mitchell said, impatient to examine Macklin for himself. "Very well. Let's step into the hall." Ferris followed them docilely. "What have you done to him?" the colonel asked straightforwardly. "We merely cured him of his headaches," Mitchell said. "How?" Mitchell did his best to explain the F-M Virus. "You mean," the Army officer said levelly "you have infected him with some kind of a disease to rot his brain?" "No, no! Could I talk to the other man, the doctor? Maybe I can make him understand." "All I want to know is why Elliot Macklin has been made as simple as if he had been kicked in the head by a mule," Colonel Carson said. "I think I can explain," Ferris interrupted. "You can?" Mitchell said. Ferris nodded. "We made a slight miscalculation. It appears as if the virus colony overcontrols the supply of posterior pituitary extract in the cerebrum. It isn't more than necessary to stop headaches. But that necessary amount of control to stop pain is too much to allow the brain cells to function properly." "Why won't they function?" Carson roared. "They don't get enough food—blood, oxygen, hemoglobin," Ferris explained. "The cerebral vessels don't contract enough to pump the blood through the brain as fast and as hard as is needed. The brain cells remain sluggish, dormant. Perhaps decaying." The colonel yelled. Mitchell groaned. He was abruptly sure Ferris was correct. The colonel drew himself to attention, fists trembling at his sides. "I'll see you hung for treason! Don't you know what Elliot Macklin means to us? Do you want those filthy Luxemburgians to reach Pluto before we do? Macklin's formula is essential to the FTL engine. You might just as well have blown up Washington, D.C. Better! The capital is replaceable. But the chances of an Elliot Macklin are very nearly once in a human race." "Just a moment," Mitchell interrupted, "we can cure Macklin." "You can ?" Carson said. For a moment Mitchell thought the man was going to clasp his hands and sink to his knees. "Certainly. We have learned to stabilize the virus colonies. We have antitoxin to combat the virus. We had always thought of it as a beneficial parasite, but we can wipe it out if necessary." "Good!" Carson clasped his hands and gave at least slightly at the knees. "Just you wait a second now, boys," Elliot Macklin said. He was leaning in the doorway, holding his pipe. "I've been listening to what you've been saying and I don't like it." "What do you mean you don't like it?" Carson demanded. He added, "Sir?" "I figure you mean to put me back like I used to be." "Yes, doctor," Mitchell said eagerly, "just as you used to be." " With my headaches, like before?" Mitchell coughed into his fist for an instant, to give him time to frame an answer. "Unfortunately, yes. Apparently if your mind functions properly once again you will have the headaches again. Our research is a dismal failure." "I wouldn't go that far," Ferris remarked cheerfully. Mitchell was about to ask his associate what he meant when he saw Macklin slowly shaking his head. "No, sir!" the mathematician said. "I shall not go back to my original state. I can remember what it was like. Always worrying, worrying, worrying." "You mean wondering," Mitchell said. Macklin nodded. "Troubled, anyway. Disturbed by every little thing. How high was up, which infinity was bigger than what infinity—say, what was an infinity anyway? All that sort of schoolboy things. It's peaceful this way. My head doesn't hurt. I've got a good-looking wife and all the money I need. I've got it made. Why worry?" Colonel Carson opened his mouth, then closed it. "That's right, Colonel. There's no use in arguing with him," Mitchell said. "It's not his decision to make," the colonel said. "He's an idiot now." "No, Colonel. As you said, he's a moron. He seems an idiot compared to his former level of intelligence but he's legally responsible. There are millions of morons running around loose in the United States. They can get married, own property, vote, even hold office. Many of them do. You can't force him into being cured.... At least, I don't think you can." "No, I can't. This is hardly a totalitarian state." The colonel looked momentarily glum that it wasn't. Mitchell looked back at Macklin. "Where did his wife get to, Colonel? I don't think that even previously he made too many personal decisions for himself. Perhaps she could influence him." "Maybe," the colonel said. "Let's find her." They found Mrs. Macklin in the dining room, her face at the picture window an attractive silhouette. She turned as the men approached. "Mrs. Macklin," the colonel began, "these gentlemen believe they can cure your husband of his present condition." "Really?" she said. "Did you speak to Elliot about that?" "Y-yes," Colonel Carson said, "but he's not himself. He refused the treatment. He wants to remain in his state of lower intelligence." She nodded. "If those are his wishes, I can't go against them." "But Mrs. Macklin!" Mitchell protested. "You will have to get a court order overruling your husband's wishes." She smoothed an eyebrow with the third finger of her right hand. "That was my original thought. But I've redecided." "Redecided!" Carson burst out almost hysterically. "Yes. I can't go against Elliot's wishes. It would be monstrous to put him back where he would suffer the hell of those headaches once again, where he never had a moment's peace from worry and pressure. He's happy now. Like a child, but happy." "Mrs. Macklin," the Army man said levelly, "if you don't help us restore your husband's mind we will be forced to get a court order declaring him incompetent." "But he is not! Legally, I mean," the woman stormed. "Maybe not. It's a borderline case. But I think any court would give us the edge where restoring the mind of Elliot Macklin was concerned. Once he's certified incompetent, authorities can rule whether Mitchell and Ferris' antitoxin treatment is the best method of restoring Dr. Macklin to sanity." "I doubt very much if the court would rule in that manner," she said. The colonel looked smug. "Why not?" "Because, Colonel, the matter of my husband's health, his very life, is involved." "There is some degree of risk in shock treatments, too. But—" "It isn't quite the same, Colonel. Elliot Macklin has a history of vascular spasm, a mild pseudostroke some years ago. Now you want to give those cerebral arteries back the ability to constrict. To paralyze. To kill. No court would give you that authority." "I suppose there's some chance of that. But without the treatment there is no chance of your husband regaining his right senses, Mrs. Macklin," Mitchell interjected. Her mouth grew petulant. "I don't care. I would rather have a live husband than a dead genius. I can take care of him this way, make him comfortable...." Carson opened his mouth and closed his fist, then relaxed. Mitchell led him back into the hall. "I'm no psychiatrist," Mitchell said, "but I think she wants Macklin stupid. Prefers it that way. She's always dominated his personal life, and now she can dominate him completely." "What is she? A monster?" the Army officer muttered. "No," Mitchell said. "She's an intelligent woman unconsciously jealous of her husband's genius." "Maybe," Carson said. "I don't know. I don't know what the hell to tell the Pentagon. I think I'll go out and get drunk." "I'll go with you," Ferris said. Mitchell glanced sharply at the little biologist. Carson squinted. "Any particular reason, doctor?" "To celebrate," Ferris said. The colonel shrugged. "That's as good a reason as any." On the street, Mitchell watched the two men go off together in bewilderment. IV Macklin was playing jacks. He didn't have a head on his shoulders and he was squatting on a great curving surface that was Spacetime, and his jacks were Earth and Pluto and the rest of the planets. And for a ball he was using a head. Not his head. Mitchell's. Both heads were initialed "M" so it was all the same. Mitchell forced himself to awaken, with some initial difficulty. He lay there, blinking the sleep out of his eyes, listening to his heart race, and then convulsively snatched the telephone receiver from the nightstand. He stabbed out a number with a vicious index finger. After a time there came a dull click and a sleepy answer. "Hello?" Elliot Macklin said. Mitchell smiled to himself. He was in luck; Macklin had answered the phone instead of his wife. "Can you speak freely, doctor?" Mitchell asked. "Of course," the mathematician said. "I can talk fine." "I mean, are you alone?" "Oh, you want to know if my wife is around. No, she's asleep. That Army doctor, Colonel Sidney, he gave her a sedative. I wouldn't let him give me anything, though." "Good boy," the biologist said. "Listen, doctor—Elliot—El, old son. I'm not against you like all the others. I don't want to make you go back to all that worrying and thinking and headaches. You believe me, don't you?" There was a slight hesitation. "Sure," Macklin said, "if you say so. Why shouldn't I believe you?" "But there was a hesitation there, El. You worried for just a second if I could have some reason for not telling you the truth." "I suppose so," Macklin said humbly. "You've found yourself worrying—thinking—about a lot of other problems since we left you, haven't you? Maybe not the same kind of scientific problem. But more personal ones, ones you didn't used to have time to think about." "If you say so." "Now, you know it's so. But how would you like to get rid of those worries just as you got rid of the others?" Mitchell asked. "I guess I'd like that," the mathematician replied. "Then come on over to my laboratory. You remember where it's at, don't you?" "No, I—yes, I guess I do. But how do I know you won't try to put me back where I was instead of helping me more?" "I couldn't do that against your wishes. That would be illegal!" "If you say so. But I don't guess I can come anyway. The Army is watching me pretty close." "That's alright," Mitchell said quickly. "You can bring along Colonel Carson." "But he won't like you fixing me up more." "But he can't stop me! Not if you want me to do it. Now listen to me—I want you to come right on over here, El." "If you say so," Macklin said uncertainly. | C. She thought they had given her husband heroin. |
How did Ned become a mummy?
A. Loy Chuk's workers wrapped Ned's body in strips of cloth to preserve it in transport.
B. The earth became a desert wasteland. All the moisture was leached from the corpse.
C. A combination of the alkali and mud his body had been soaked in. Also, the years of dryness after the world became a desert.
D. The body had been devoid of moisture for a million years.
| THE ETERNAL WALL By RAYMOND Z. GALLUN A scream of brakes, the splash into icy waters, a long descent into alkaline depths ... it was death. But Ned Vince lived again—a million years later! "See you in half an hour, Betty," said Ned Vince over the party telephone. "We'll be out at the Silver Basket before ten-thirty...." Ned Vince was eager for the company of the girl he loved. That was why he was in a hurry to get to the neighboring town of Hurley, where she lived. His old car rattled and roared as he swung it recklessly around Pit Bend. There was where Death tapped him on the shoulder. Another car leaped suddenly into view, its lights glaring blindingly past a high, up-jutting mass of Jurassic rock at the turn of the road. Dazzled, and befuddled by his own rash speed, Ned Vince had only swift young reflexes to rely on to avoid a fearful, telescoping collision. He flicked his wheel smoothly to the right; but the County Highway Commission hadn't yet tarred the traffic-loosened gravel at the Bend. An incredible science, millions of years old, lay in the minds of these creatures. Ned could scarcely have chosen a worse place to start sliding and spinning. His car hit the white-painted wooden rail sideways, crashed through, tumbled down a steep slope, struck a huge boulder, bounced up a little, and arced outward, falling as gracefully as a swan-diver toward the inky waters of the Pit, fifty feet beneath.... Ned Vince was still dimly conscious when that black, quiet pool geysered around him in a mighty splash. He had only a dazing welt on his forehead, and a gag of terror in his throat. Movement was slower now, as he began to sink, trapped inside his wrecked car. Nothing that he could imagine could mean doom more certainly than this. The Pit was a tremendously deep pocket in the ground, spring-fed. The edges of that almost bottomless pool were caked with a rim of white—for the water, on which dead birds so often floated, was surcharged with alkali. As that heavy, natronous liquid rushed up through the openings and cracks beneath his feet, Ned Vince knew that his friends and his family would never see his body again, lost beyond recovery in this abyss. The car was deeply submerged. The light had blinked out on the dash-panel, leaving Ned in absolute darkness. A flood rushed in at the shattered window. He clawed at the door, trying to open it, but it was jammed in the crash-bent frame, and he couldn't fight against the force of that incoming water. The welt, left by the blow he had received on his forehead, put a thickening mist over his brain, so that he could not think clearly. Presently, when he could no longer hold his breath, bitter liquid was sucked into his lungs. His last thoughts were those of a drowning man. The machine-shop he and his dad had had in Harwich. Betty Moore, with the smiling Irish eyes—like in the song. Betty and he had planned to go to the State University this Fall. They'd planned to be married sometime.... Goodbye, Betty ... The ripples that had ruffled the surface waters in the Pit, quieted again to glassy smoothness. The eternal stars shone calmly. The geologic Dakota hills, which might have seen the dinosaurs, still bulked along the highway. Time, the Brother of Death, and the Father of Change, seemed to wait.... "Kaalleee! Tik!... Tik, tik, tik!... Kaalleee!..." The excited cry, which no human throat could quite have duplicated accurately, arose thinly from the depths of a powder-dry gulch, water-scarred from an inconceivable antiquity. The noon-day Sun was red and huge. The air was tenuous, dehydrated, chill. "Kaalleee!... Tik, tik, tik!..." At first there was only one voice uttering those weird, triumphant sounds. Then other vocal organs took up that trilling wail, and those short, sharp chuckles of eagerness. Other questioning, wondering notes mixed with the cadence. Lacking qualities identifiable as human, the disturbance was still like the babble of a group of workmen who have discovered something remarkable. The desolate expanse around the gulch, was all but without motion. The icy breeze tore tiny puffs of dust from grotesque, angling drifts of soil, nearly waterless for eons. Patches of drab lichen grew here and there on the up-jutting rocks, but in the desert itself, no other life was visible. Even the hills had sagged away, flattened by incalculable ages of erosion. At a mile distance, a crumbling heap of rubble arose. Once it had been a building. A gigantic, jagged mass of detritus slanted upward from its crest—red debris that had once been steel. A launching catapult for the last space ships built by the gods in exodus, perhaps it was—half a million years ago. Man was gone from the Earth. Glacial ages, war, decadence, disease, and a final scattering of those ultimate superhumans to newer worlds in other solar systems, had done that. "Kaalleee!... Tik, tik, tik!..." The sounds were not human. They were more like the chatter and wail of small desert animals. But there was a seeming paradox here in the depths of that gulch, too. The glint of metal, sharp and burnished. The flat, streamlined bulk of a flying machine, shiny and new. The bell-like muzzle of a strange excavator-apparatus, which seemed to depend on a blast of atoms to clear away rock and soil. Thus the gulch had been cleared of the accumulated rubbish of antiquity. Man, it seemed, had a successor, as ruler of the Earth. Loy Chuk had flown his geological expedition out from the far lowlands to the east, out from the city of Kar-Rah. And he was very happy now—flushed with a vast and unlooked-for success. He crouched there on his haunches, at the dry bottom of the Pit. The breeze rumpled his long, brown fur. He wasn't very different in appearance from his ancestors. A foot tall, perhaps, as he squatted there in that antique stance of his kind. His tail was short and furred, his undersides creamy. White whiskers spread around his inquisitive, pink-tipped snout. But his cranium bulged up and forward between shrewd, beady eyes, betraying the slow heritage of time, of survival of the fittest, of evolution. He could think and dream and invent, and the civilization of his kind was already far beyond that of the ancient Twentieth Century. Loy Chuk and his fellow workers were gathered, tense and gleeful, around the things their digging had exposed to the daylight. There was a gob of junk—scarcely more than an irregular formation of flaky rust. But imbedded in it was a huddled form, brown and hard as old wood. The dry mud that had encased it like an airtight coffin, had by now been chipped away by the tiny investigators; but soiled clothing still clung to it, after perhaps a million years. Metal had gone into decay—yes. But not this body. The answer to this was simple—alkali. A mineral saturation that had held time and change in stasis. A perfect preservative for organic tissue, aided probably during most of those passing eras by desert dryness. The Dakotas had turned arid very swiftly. This body was not a mere fossil. It was a mummy. "Kaalleee!" Man, that meant. Not the star-conquering demi-gods, but the ancestral stock that had built the first machines on Earth, and in the early Twenty-first Century, the first interplanetary rockets. No wonder Loy Chuk and his co-workers were happy in their paleontological enthusiasm! A strange accident, happening in a legendary antiquity, had aided them in their quest for knowledge. At last Loy Chuk gave a soft, chirping signal. The chant of triumph ended, while instruments flicked in his tiny hands. The final instrument he used to test the mummy, looked like a miniature stereoscope, with complicated details. He held it over his eyes. On the tiny screen within, through the agency of focused X-rays, he saw magnified images of the internal organs of this ancient human corpse. What his probing gaze revealed to him, made his pleasure even greater than before. In twittering, chattering sounds, he communicated his further knowledge to his henchmen. Though devoid of moisture, the mummy was perfectly preserved, even to its brain cells! Medical and biological sciences were far advanced among Loy Chuk's kind. Perhaps, by the application of principles long known to them, this long-dead body could be made to live again! It might move, speak, remember its past! What a marvelous subject for study it would make, back there in the museums of Kar-Rah! "Tik, tik, tik!..." But Loy silenced this fresh, eager chattering with a command. Work was always more substantial than cheering. With infinite care—small, sharp hand-tools were used, now—the mummy of Ned Vince was disengaged from the worthless rust of his primitive automobile. With infinite care it was crated in a metal case, and hauled into the flying machine. Flashing flame, the latter arose, bearing the entire hundred members of the expedition. The craft shot eastward at bullet-like speed. The spreading continental plateau of North America seemed to crawl backward, beneath. A tremendous sand desert, marked with low, washed-down mountains, and the vague, angular, geometric mounds of human cities that were gone forever. Beyond the eastern rim of the continent, the plain dipped downward steeply. The white of dried salt was on the hills, but there was a little green growth here, too. The dead sea-bottom of the vanished Atlantic was not as dead as the highlands. Far out in a deep valley, Kar-Rah, the city of the rodents, came into view—a crystalline maze of low, bubble-like structures, glinting in the red sunshine. But this was only its surface aspect. Loy Chuk's people had built their homes mostly underground, since the beginning of their foggy evolution. Besides, in this latter day, the nights were very cold, the shelter of subterranean passages and rooms was welcome. The mummy was taken to Loy Chuk's laboratory, a short distance below the surface. Here at once, the scientist began his work. The body of the ancient man was put in a large vat. Fluids submerged it, slowly soaking from that hardened flesh the alkali that had preserved it for so long. The fluid was changed often, until woody muscles and other tissues became pliable once more. Then the more delicate processes began. Still submerged in liquid, the corpse was submitted to a flow of restorative energy, passing between complicated electrodes. The cells of antique flesh and brain gradually took on a chemical composition nearer to that of the life that they had once known. At last the final liquid was drained away, and the mummy lay there, a mummy no more, but a pale, silent figure in its tatters of clothing. Loy Chuk put an odd, metal-fabric helmet on its head, and a second, much smaller helmet on his own. Connected with this arrangement, was a black box of many uses. For hours he worked with his apparatus, studying, and guiding the recording instruments. The time passed swiftly. At last, eager and ready for whatever might happen now, Loy Chuk pushed another switch. With a cold, rosy flare, energy blazed around that moveless form. For Ned Vince, timeless eternity ended like a gradual fading mist. When he could see clearly again, he experienced that inevitable shock of vast change around him. Though it had been dehydrated, his brain had been kept perfectly intact through the ages, and now it was restored. So his memories were as vivid as yesterday. Yet, through that crystalline vat in which he lay, he could see a broad, low room, in which he could barely have stood erect. He saw instruments and equipment whose weird shapes suggested alienness, and knowledge beyond the era he had known! The walls were lavender and phosphorescent. Fossil bone-fragments were mounted in shallow cases. Dinosaur bones, some of them seemed, from their size. But there was a complete skeleton of a dog, too, and the skeleton of a man, and a second man-skeleton that was not quite human. Its neck-vertebrae were very thick and solid, its shoulders were wide, and its skull was gigantic. All this weirdness had a violent effect on Ned Vince—a sudden, nostalgic panic. Something was fearfully wrong! The nervous terror of the unknown was on him. Feeble and dizzy after his weird resurrection, which he could not understand, remembering as he did that moment of sinking to certain death in the pool at Pit Bend, he caught the edge of the transparent vat, and pulled himself to a sitting posture. There was a muffled murmur around him, as of some vast, un-Earthly metropolis. "Take it easy, Ned Vince...." The words themselves, and the way they were assembled, were old, familiar friends. But the tone was wrong. It was high, shrill, parrot-like, and mechanical. Ned's gaze searched for the source of the voice—located the black box just outside of his crystal vat. From that box the voice seemed to have originated. Before it crouched a small, brownish animal with a bulging head. The animal's tiny-fingered paws—hands they were, really—were touching rows of keys. To Ned Vince, it was all utterly insane and incomprehensible. A rodent, looking like a prairie dog, a little; but plainly possessing a high order of intelligence. And a voice whose soothingly familiar words were more repugnant somehow, simply because they could never belong in a place as eerie as this. Ned Vince did not know how Loy Chuk had probed his brain, with the aid of a pair of helmets, and the black box apparatus. He did not know that in the latter, his language, taken from his own revitalized mind, was recorded, and that Loy Chuk had only to press certain buttons to make the instrument express his thoughts in common, long-dead English. Loy, whose vocal organs were not human, would have had great difficulty speaking English words, anyway. Ned's dark hair was wildly awry. His gaunt, young face held befuddled terror. He gasped in the thin atmosphere. "I've gone nuts," he pronounced with a curious calm. "Stark—starin'—nuts...." Loy's box, with its recorded English words and its sonic detectors, could translate for its master, too. As the man spoke, Loy read the illuminated symbols in his own language, flashed on a frosted crystal plate before him. Thus he knew what Ned Vince was saying. Loy Chuk pressed more keys, and the box reproduced his answer: "No, Ned, not nuts. Not a bit of it! There are just a lot of things that you've got to get used to, that's all. You drowned about a million years ago. I discovered your body. I brought you back to life. We have science that can do that. I'm Loy Chuk...." It took only a moment for the box to tell the full story in clear, bold, friendly terms. Thus Loy sought, with calm, human logic, to make his charge feel at home. Probably, though, he was a fool, to suppose that he could succeed, thus. Vince started to mutter, struggling desperately to reason it out. "A prairie dog," he said. "Speaking to me. One million years. Evolution. The scientists say that people grew up from fishes in the sea. Prairie dogs are smart. So maybe super-prairie-dogs could come from them. A lot easier than men from fish...." It was all sound logic. Even Ned Vince knew that. Still, his mind, tuned to ordinary, simple things, couldn't quite realize all the vast things that had happened to himself, and to the world. The scope of it all was too staggeringly big. One million years. God!... Ned Vince made a last effort to control himself. His knuckles tightened on the edge of the vat. "I don't know what you've been talking about," he grated wildly. "But I want to get out of here! I want to go back where I came from! Do you understand—whoever, or whatever you are?" Loy Chuk pressed more keys. "But you can't go back to the Twentieth Century," said the box. "Nor is there any better place for you to be now, than Kar-Rah. You are the only man left on Earth. Those men that exist in other star systems are not really your kind anymore, though their forefathers originated on this planet. They have gone far beyond you in evolution. To them you would be only a senseless curiosity. You are much better off with my people—our minds are much more like yours. We will take care of you, and make you comfortable...." But Ned Vince wasn't listening, now. "You are the only man left on Earth." That had been enough for him to hear. He didn't more than half believe it. His mind was too confused for conviction about anything. Everything he saw and felt and heard might be some kind of nightmare. But then it might all be real instead, and that was abysmal horror. Ned was no coward—death and danger of any ordinary Earthly kind, he could have faced bravely. But the loneliness here, and the utter strangeness, were hideous like being stranded alone on another world! His heart was pounding heavily, and his eyes were wide. He looked across this eerie room. There was a ramp there at the other side, leading upward instead of a stairway. Fierce impulse to escape this nameless lair, to try to learn the facts for himself, possessed him. He bounded out of the vat, and with head down, dashed for the ramp. He had to go most of the way on his hands and knees, for the up-slanting passage was low. Excited animal chucklings around him, and the occasional touch of a furry body, hurried his feverish scrambling. But he emerged at last at the surface. He stood there panting in that frigid, rarefied air. It was night. The Moon was a gigantic, pock-marked bulk. The constellations were unrecognizable. The rodent city was a glowing expanse of shallow, crystalline domes, set among odd, scrub trees and bushes. The crags loomed on all sides, all their jaggedness lost after a million years of erosion under an ocean that was gone. In that ghastly moonlight, the ground glistened with dry salt. "Well, I guess it's all true, huh?" Ned Vince muttered in a flat tone. Behind him he heard an excited, squeaky chattering. Rodents in pursuit. Looking back, he saw the pinpoint gleams of countless little eyes. Yes, he might as well be an exile on another planet—so changed had the Earth become. A wave of intolerable homesickness came over him as he sensed the distances of time that had passed—those inconceivable eons, separating himself from his friends, from Betty, from almost everything that was familiar. He started to run, away from those glittering rodent eyes. He sensed death in that cold sea-bottom, but what of it? What reason did he have left to live? He'd be only a museum piece here, a thing to be caged and studied.... Prison or a madhouse would be far better. He tried to get hold of his courage. But what was there to inspire it? Nothing! He laughed harshly as he ran, welcoming that bitter, killing cold. Nostalgia had him in its clutch, and there was no answer in his hell-world, lost beyond the barrier of the years.... Loy Chuk and his followers presently came upon Ned Vince's unconscious form, a mile from the city of Kar-Rah. In a flying machine they took him back, and applied stimulants. He came to, in the same laboratory room as before. But he was firmly strapped to a low platform this time, so that he could not escape again. There he lay, helpless, until presently an idea occurred to him. It gave him a few crumbs of hope. "Hey, somebody!" he called. "You'd better get some rest, Ned Vince," came the answer from the black box. It was Loy Chuk speaking again. "But listen!" Ned protested. "You know a lot more than we did in the Twentieth Century. And—well—there's that thing called time-travel, that I used to read about. Maybe you know how to make it work! Maybe you could send me back to my own time after all!" Little Loy Chuk was in a black, discouraged mood, himself. He could understand the utter, sick dejection of this giant from the past, lost from his own kind. Probably insanity looming. In far less extreme circumstances than this, death from homesickness had come. Loy Chuk was a scientist. In common with all real scientists, regardless of the species from which they spring, he loved the subjects of his researches. He wanted this ancient man to live and to be happy. Or this creature would be of scant value for study. So Loy considered carefully what Ned Vince had suggested. Time-travel. Almost a legend. An assault upon an intangible wall that had baffled far keener wits than Loy's. But he was bent, now, on the well-being of this anachronism he had so miraculously resurrected—this human, this Kaalleee.... Loy jabbed buttons on the black box. "Yes, Ned Vince," said the sonic apparatus. "Time-travel. Perhaps that is the only thing to do—to send you back to your own period of history. For I see that you will never be yourself, here. It will be hard to accomplish, but we'll try. Now I shall put you under an anesthetic...." Ned felt better immediately, for there was real hope now, where there had been none before. Maybe he'd be back in his home-town of Harwich again. Maybe he'd see the old machine-shop, there. And the trees greening out in Spring. Maybe he'd be seeing Betty Moore in Hurley, soon.... Ned relaxed, as a tiny hypo-needle bit into his arm.... As soon as Ned Vince passed into unconsciousness, Loy Chuk went to work once more, using that pair of brain-helmets again, exploring carefully the man's mind. After hours of research, he proceeded to prepare his plans. The government of Kar-Rah was a scientific oligarchy, of which Loy was a prime member. It would be easy to get the help he needed. A horde of small, grey-furred beings and their machines, toiled for many days. Ned Vince's mind swam gradually out of the blur that had enveloped it. He was wandering aimlessly about in a familiar room. The girders of the roof above were of red-painted steel. His tool-benches were there, greasy and littered with metal filings, just as they had always been. He had a tractor to repair, and a seed-drill. Outside of the machine-shop, the old, familiar yellow sun was shining. Across the street was the small brown house, where he lived. With a sudden startlement, he saw Betty Moore in the doorway. She wore a blue dress, and a mischievous smile curved her lips. As though she had succeeded in creeping up on him, for a surprise. "Why, Ned," she chuckled. "You look as though you've been dreaming, and just woke up!" He grimaced ruefully as she approached. With a kind of fierce gratitude, he took her in his arms. Yes, she was just like always. "I guess I was dreaming, Betty," he whispered, feeling that mighty sense of relief. "I must have fallen asleep at the bench, here, and had a nightmare. I thought I had an accident at Pit Bend—and that a lot of worse things happened.... But it wasn't true ..." Ned Vince's mind, over which there was still an elusive fog that he did not try to shake off, accepted apparent facts simply. He did not know anything about the invisible radiations beating down upon him, soothing and dimming his brain, so that it would never question or doubt, or observe too closely the incongruous circumstances that must often appear. The lack of traffic in the street without, for instance—and the lack of people besides himself and Betty. He didn't know that this machine-shop was built from his own memories of the original. He didn't know that this Betty was of the same origin—a miraculous fabrication of metal and energy-units and soft plastic. The trees outside were only lantern-slide illusions. It was all built inside a great, opaque dome. But there were hidden television systems, too. Thus Loy Chuk's kind could study this ancient man—this Kaalleee. Thus, their motives were mostly selfish. Loy, though, was not observing, now. He had wandered far out into cold, sad sea-bottom, to ponder. He squeaked and chatted to himself, contemplating the magnificent, inexorable march of the ages. He remembered the ancient ruins, left by the final supermen. "The Kaalleee believes himself home," Loy was thinking. "He will survive and be happy. But there was no other way. Time is an Eternal Wall. Our archeological researches among the cities of the supermen show the truth. Even they, who once ruled Earth, never escaped from the present by so much as an instant...." THE END PRINTED IN U. S. A. Transcriber's Note: This etext was produced from Amazing Stories April 1956 and was first published in Amazing Stories November 1942. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed. Minor spelling and typographical errors have been corrected without note. | C. A combination of the alkali and mud his body had been soaked in. Also, the years of dryness after the world became a desert. |
What programming language is target language? | ### Introduction
Removing computer-human language barrier is an inevitable advancement researchers are thriving to achieve for decades. One of the stages of this advancement will be coding through natural human language instead of traditional programming language. On naturalness of computer programming D. Knuth said, “Let us change our traditional attitude to the construction of programs: Instead of imagining that our main task is to instruct a computer what to do, let us concentrate rather on explaining to human beings what we want a computer to do.”BIBREF0. Unfortunately, learning programming language is still necessary to instruct it. Researchers and developers are working to overcome this human-machine language barrier. Multiple branches exists to solve this challenge (i.e. inter-conversion of different programming language to have universally connected programming languages). Automatic code generation through natural language is not a new concept in computer science studies. However, it is difficult to create such tool due to these following three reasons– Programming languages are diverse An individual person expresses logical statements differently than other Natural Language Processing (NLP) of programming statements is challenging since both human and programming language evolve over time In this paper, a neural approach to translate pseudo-code or algorithm like human language expression into programming language code is proposed. ### Problem Description
Code repositories (i.e. Git, SVN) flourished in the last decade producing big data of code allowing data scientists to perform machine learning on these data. In 2017, Allamanis M et al. published a survey in which they presented the state-of-the-art of the research areas where machine learning is changing the way programmers code during software engineering and development process BIBREF1. This paper discusses what are the restricting factors of developing such text-to-code conversion method and what problems need to be solved– ### Problem Description ::: Programming Language Diversity
According to the sources, there are more than a thousand actively maintained programming languages, which signifies the diversity of these language . These languages were created to achieve different purpose and use different syntaxes. Low-level languages such as assembly languages are easier to express in human language because of the low or no abstraction at all whereas high-level, or Object-Oriented Programing (OOP) languages are more diversified in syntax and expression, which is challenging to bring into a unified human language structure. Nonetheless, portability and transparency between different programming languages also remains a challenge and an open research area. George D. et al. tried to overcome this problem through XML mapping BIBREF2. They tried to convert codes from C++ to Java using XML mapping as an intermediate language. However, the authors encountered challenges to support different features of both languages. ### Problem Description ::: Human Language Factor
One of the motivations behind this paper is - as long as it is about programming, there is a finite and small set of expression which is used in human vocabulary. For instance, programmers express a for-loop in a very few specific ways BIBREF3. Variable declaration and value assignment expressions are also limited in nature. Although all codes are executable, human representation through text may not due to the semantic brittleness of code. Since high-level languages have a wide range of syntax, programmers use different linguistic expressions to explain those. For instance, small changes like swapping function arguments can significantly change the meaning of the code. Hence the challenge remains in processing human language to understand it properly which brings us to the next problem- ### Problem Description ::: NLP of statements
Although there is a finite set of expressions for each programming statements, it is a challenge to extract information from the statements of the code accurately. Semantic analysis of linguistic expression plays an important role in this information extraction. For instance, in case of a loop, what is the initial value? What is the step value? When will the loop terminate? Mihalcea R. et al. has achieved a variable success rate of 70-80% in producing code just from the problem statement expressed in human natural language BIBREF3. They focused solely on the detection of step and loops in their research. Another research group from MIT, Lei et al. use a semantic learning model for text to detect the inputs. The model produces a parser in C++ which can successfully parse more than 70% of the textual description of input BIBREF4. The test dataset and model was initially tested and targeted against ACM-ICPC participantsínputs which contains diverse and sometimes complex input instructions. A recent survey from Allamanis M. et al. presented the state-of-the-art on the area of naturalness of programming BIBREF1. A number of research works have been conducted on text-to-code or code-to-text area in recent years. In 2015, Oda et al. proposed a way to translate each line of Python code into natural language pseudocode using Statistical Machine Learning Technique (SMT) framework BIBREF5 was used. This translation framework was able to - it can successfully translate the code to natural language pseudo coded text in both English and Japanese. In the same year, Chris Q. et al. mapped natural language with simple if-this-then-that logical rules BIBREF6. Tihomir G. and Viktor K. developed an Integrated Development Environment (IDE) integrated code assistant tool anyCode for Java which can search, import and call function just by typing desired functionality through text BIBREF7. They have used model and mapping framework between function signatures and utilized resources like WordNet, Java Corpus, relational mapping to process text online and offline. Recently in 2017, P. Yin and G. Neubig proposed a semantic parser which generates code through its neural model BIBREF8. They formulated a grammatical model which works as a skeleton for neural network training. The grammatical rules are defined based on the various generalized structure of the statements in the programming language. ### Proposed Methodology
The use of machine learning techniques such as SMT proved to be at most 75% successful in converting human text to executable code. BIBREF9. A programming language is just like a language with less vocabulary compared to a typical human language. For instance, the code vocabulary of the training dataset was 8814 (including variable, function, class names), whereas the English vocabulary to express the same code was 13659 in total. Here, programming language is considered just like another human language and widely used SMT techniques have been applied. ### Proposed Methodology ::: Statistical Machine Translation
SMT techniques are widely used in Natural Language Processing (NLP). SMT plays a significant role in translation from one language to another, especially in lexical and grammatical rule extraction. In SMT, bilingual grammatical structures are automatically formed by statistical approaches instead of explicitly providing a grammatical model. This reduces months and years of work which requires significant collaboration between bi-lingual linguistics. Here, a neural network based machine translation model is used to translate regular text into programming code. ### Proposed Methodology ::: Statistical Machine Translation ::: Data Preparation
SMT techniques require a parallel corpus in thr source and thr target language. A text-code parallel corpus similar to Fig. FIGREF12 is used in training. This parallel corpus has 18805 aligned data in it . In source data, the expression of each line code is written in the English language. In target data, the code is written in Python programming language. ### Proposed Methodology ::: Statistical Machine Translation ::: Vocabulary Generation
To train the neural model, the texts should be converted to a computational entity. To do that, two separate vocabulary files are created - one for the source texts and another for the code. Vocabulary generation is done by tokenization of words. Afterwards, the words are put into their contextual vector space using the popular word2vec BIBREF10 method to make the words computational. ### Proposed Methodology ::: Statistical Machine Translation ::: Neural Model Training
In order to train the translation model between text-to-code an open source Neural Machine Translation (NMT) - OpenNMT implementation is utilized BIBREF11. PyTorch is used as Neural Network coding framework. For training, three types of Recurrent Neural Network (RNN) layers are used – an encoder layer, a decoder layer and an output layer. These layers together form a LSTM model. LSTM is typically used in seq2seq translation. In Fig. FIGREF13, the neural model architecture is demonstrated. The diagram shows how it takes the source and target text as input and uses it for training. Vector representation of tokenized source and target text are fed into the model. Each token of the source text is passed into an encoder cell. Target text tokens are passed into a decoder cell. Encoder cells are part of the encoder RNN layer and decoder cells are part of the decoder RNN layer. End of the input sequence is marked by a $<$eos$>$ token. Upon getting the $<$eos$>$ token, the final cell state of encoder layer initiate the output layer sequence. At each target cell state, attention is applied with the encoder RNN state and combined with the current hidden state to produce the prediction of next target token. This predictions are then fed back to the target RNN. Attention mechanism helps us to overcome the fixed length restriction of encoder-decoder sequence and allows us to process variable length between input and output sequence. Attention uses encoder state and pass it to the decoder cell to give particular attention to the start of an output layer sequence. The encoder uses an initial state to tell the decoder what it is supposed to generate. Effectively, the decoder learns to generate target tokens, conditioned on the input sequence. Sigmoidal optimization is used to optimize the prediction. ### Result Analysis
Training parallel corpus had 18805 lines of annotated code in it. The training model is executed several times with different training parameters. During the final training process, 500 validation data is used to generate the recurrent neural model, which is 3% of the training data. We run the training with epoch value of 10 with a batch size of 64. After finishing the training, the accuracy of the generated model using validation data from the source corpus was 74.40% (Fig. FIGREF17). Although the generated code is incoherent and often predict wrong code token, this is expected because of the limited amount of training data. LSTM generally requires a more extensive set of data (100k+ in such scenario) to build a more accurate model. The incoherence can be resolved by incorporating coding syntax tree model in future. For instance– "define the method tzname with 2 arguments: self and dt." is translated into– def __init__ ( self , regex ) :. The translator is successfully generating the whole codeline automatically but missing the noun part (parameter and function name) part of the syntax. ### Conclusion & Future Works
The main advantage of translating to a programming language is - it has a concrete and strict lexical and grammatical structure which human languages lack. The aim of this paper was to make the text-to-code framework work for general purpose programming language, primarily Python. In later phase, phrase-based word embedding can be incorporated for improved vocabulary mapping. To get more accurate target code for each line, Abstract Syntax Tree(AST) can be beneficial. The contribution of this research is a machine learning model which can turn the human expression into coding expressions. This paper also discusses available methods which convert natural language to programming language successfully in fixed or tightly bounded linguistic paradigm. Approaching this problem using machine learning will give us the opportunity to explore the possibility of unified programming interface as well in the future. ### Acknowledgment
We would like to thank Dr. Khandaker Tabin Hasan, Head of the Depertment of Computer Science, American International University-Bangladesh for his inspiration and encouragement in all of our research works. Also, thanks to Future Technology Conference - 2019 committee for partially supporting us to join the conference and one of our colleague - Faheem Abrar, Software Developer for his thorough review and comments on this research work and supporting us by providing fund. Fig. 1. Text-Code bi-lingual corpus Fig. 2. Neural training model architecture of Text-To-Code Fig. 3. Accuracy gain in progress of training the RNN | Python |
What data augmentation techniques are used? | ### Introduction
Virtual assistants help users accomplish tasks including but not limited to finding flights, booking restaurants, by providing a natural language interface to services and APIs on the web. Large-scale assistants like Google Assistant, Amazon Alexa, Apple Siri, Microsoft Cortana etc. need to support a large and constantly increasing number of services, over a wide variety of domains. Consequently, recent work has focused on scalable dialogue systems that can handle tasks across multiple application domains. Data-driven deep learning based approaches for multi-domain modeling have shown promise, both for end-to-end and modular systems involving dialogue state tracking and policy learning. This line of work has been facilitated by the release of multi-domain dialogue corpora such as MultiWOZ BIBREF0, Taskmaster-1 BIBREF1, M2M BIBREF2 and FRAMES BIBREF3. However, building large-scale assistants, as opposed to dialogue systems managing a few APIs, poses a new set of challenges. Apart from the handling a very large variety of domains, such systems need to support heterogeneous services or APIs with possibly overlapping functionality. It should also offer an efficient way of supporting new APIs or services, while requiring little or no additional training data. Furthermore, to reduce maintenance workload and accommodate future growth, such assistants need to be robust to changes in the API's interface or addition of new slot values. Such changes shouldn't require collection of additional training data or retraining the model. The Schema-Guided Dialogue State Tracking task at the Eighth Dialogue System Technology Challenge explores the aforementioned challenges in context of dialogue state tracking. In a task-oriented dialogue, the dialogue state is a summary of the entire conversation till the current turn. The dialogue state is used to invoke APIs with appropriate parameters as specified by the user over the dialogue history. It is also used by the assistant to generate the next actions to continue the dialogue. DST, therefore, is a core component of virtual assistants. In this task, participants are required to develop innovative approaches to multi-domain dialogue state tracking, with a focus on data-efficient joint modeling across APIs and zero-shot generalization to new APIs. The task is based on the Schema-Guided Dialogue (SGD) dataset, which, to the best of our knowledge, is the largest publicly available corpus of annotated task-oriented dialogues. With over 16000 dialogues in the training set spanning 26 APIs over 16 domains, it exceeds the existing dialogue corpora in scale. SGD is the first dataset to allow multiple APIs with overlapping functionality within each domain. To adequately test generalization in zero-shot settings, the evaluation sets contain unseen services and domains. The dataset is designed to serve as an effective testbed for intent prediction, slot filling, state tracking and language generation, among other tasks in large-scale virtual assistants. ### Related Work
Dialogue systems have constituted an active area of research for the past few decades. The advent of commercial personal assistants has provided further impetus to dialogue systems research. As virtual assistants incorporate diverse domains, zero-shot modeling BIBREF4, BIBREF5, BIBREF6, domain adaptation and transfer learning techniques BIBREF7, BIBREF8, BIBREF9 have been explored to support new domains in a data efficient manner. Deep learning based approaches to DST have recently gained popularity. Some of these approaches estimate the dialogue state as a distribution over all possible slot-values BIBREF10, BIBREF11 or individually score all slot-value combinations BIBREF12, BIBREF13. Such approaches are, however, hard to scale to real-world virtual assistants, where the set of possible values for certain slots may be very large (date, time or restaurant name) and even dynamic (movie or event name). Other approaches utilizing a dynamic vocabulary of slot values BIBREF14, BIBREF15 still do not allow zero-shot generalization to new services and APIs BIBREF16, since they use schema elements i.e. intents and slots as fixed class labels. Although such systems are capable of parsing the dialogue semantics in terms of these fixed intent labels, they lack understanding of the semantics of these labels. For instance, for the user utterance “I want to buy tickets for a movie.", such models can predict BuyMovieTickets as the correct intent based on patterns observed in the training data, but don't model either its association with the real world action of buying movie tickets, or its similarity to the action of buying concert or theatre tickets. Furthermore, because of their dependence on a fixed schema, such models are not robust to changes in the schema, and need to be retrained as new slots or intents are added. Use of domain-specific parameters renders some approaches unsuitable for zero-shot application. ### Task
The primary task of this challenge is to develop multi-domain models for DST suitable for the scale and complexity of large scale virtual assistants. Supporting a wide variety of APIs or services with possibly overlapping functionality is an important requirement of such assistants. A common approach to do this involves defining a large master schema that lists all intents and slots supported by the assistant. Each service either adopts this master schema for the representation of the underlying data, or provides logic to translate between its own schema and the master schema. The first approach involving adoption of the master schema is not ideal if a service wishes to integrate with multiple assistants, since each of the assistants could have their own master schema. The second approach involves definition of logic for translation between master schema and the service's schema, which increases the maintenance workload. Furthermore, it is difficult to develop a master schema catering to all possible use cases. Additionally, while there are many similar concepts across services that can be jointly modeled, for example, the similarities in logic for querying or specifying the number of movie tickets, flight tickets or concert tickets, the master schema approach does not facilitate joint modeling of such concepts, unless an explicit mapping between them is manually defined. To address these limitations, we propose a schema-guided approach, which eliminates the need for a master schema. ### Task ::: Schema-Guided Approach
Under the Schema-Guided approach, each service provides a schema listing the supported slots and intents along with their natural language descriptions (Figure FIGREF2 shows an example). The dialogue annotations are guided by the schema of the underlying service or API, as shown in Figure FIGREF3. In this example, the departure and arrival cities are captured by analogously functioning but differently named slots in both schemas. Furthermore, values for the number_stops and direct_only slots highlight idiosyncrasies between services interpreting the same concept. The natural language descriptions present in the schema are used to obtain a semantic representation of intents and slots. The assistant employs a single unified model containing no domain or service specific parameters to make predictions conditioned on these schema elements. Using a single model facilitates representation and transfer of common knowledge across related concepts in different services. Since the model utilizes semantic representation of schema elements as input, it can interface with unseen services or APIs on which it has not been trained. It is also robust to changes like the addition of new intents or slots to the service. In addition, the participants are allowed to use any external datasets or resources to bootstrap their models. ### Dataset
As shown in Table TABREF9, our Schema-Guided Dialogue (SGD) dataset exceeds other datasets in most of the metrics at scale. The especially larger number of domains, slots, and slot values, and the presence of multiple services per domain, are representative of these scale-related challenges. Furthermore, our evaluation sets contain many services, and consequently slots, which are not present in the training set, to help evaluate model performance on unseen services. ### Dataset ::: Data Representation
The dataset consists of conversations between a virtual assistant and a user. Each conversation can span multiple services across various domains. The dialogue is represented as a sequence of turns, each containing a user or system utterance. The annotations for each turn are grouped into frames, where each frame corresponds to a single service. The annotations for user turns include the active intent, the dialogue state and slot spans for the different slots values mentioned in the turn. For system turns, we have the system actions representing the semantics of the system utterance. Each system action is represented using a dialogue act with optional parameters. In addition to the dialogues, for each service used in the dataset, a normalized representation of the interface exposed is provided as the schema. The schema contains details like the name of the service, the list of tasks supported by the service (intents) and the attributes of the entities used by the service (slots). The schema also contains natural language descriptions of the service, intents and slots which can be used for developing models which can condition their predictions on the schema. ### Dataset ::: Comparison With Other Datasets
To reflect the constraints present in real-world services and APIs, we impose a few constraints on the data. Our dataset does not expose the set of all possible values for certain slots. Having such a list is impractical for slots like date or time because they have infinitely many possible values or for slots like movie or song names, for which new values are periodically added. Such slots are specifically identified as non-categorical slots. In our evaluation sets, we ensured the presence of a significant number of values which were not previously seen in the training set to evaluate the performance of models on unseen values. Some slots like gender, number of people, etc. are classified as categorical and we provide a list of all possible values for them. However, these values are assumed to be not consistent across services. E.g., different services may use (`male', `female'), (`M', `F') or (`he', `she') as possible values for gender slot. Real-world services can only be invoked with certain slot combinations: e.g. most restaurant reservation APIs do not let the user search for restaurants by date without specifying a location. Although this constraint has no implications on the dialogue state tracking task, it restricts the possible conversational flows. Hence, to prevent flows not supported by actual services, we restrict services to be called with a list of slot combinations. The different service calls supported by a service are listed as intents with each intent specifying a list of required slots. The intent cannot be called without providing values for these required slots. Each intent also contains a list of optional slots with default values which can be overridden by the user. In our dataset, we also have multiple services per domain with overlapping functionality. The intents across these services are similar but differ in terms of intent names, intent arguments, slot names, etc. In some cases, there is no one to one mapping between slot names (e.g., the num_stops and direct_only slots in Figure FIGREF3). With an ever increasing number of services and service providers, we believe that having multiple similar services per domain is much closer to the situation faced by virtual assistants than having one unique service per domain. ### Dataset ::: Data Collection And Dataset Analysis
Our data collection setup uses a dialogue simulator to generate dialogue outlines first and then paraphrase them to obtain natural utterances. Using a dialogue simulator offers us multiple advantages. First, it ensures the coverage of a large variety of dialogue flows by filtering out similar flows in the simulation phase, thus creating a much diverse dataset. Second, simulated dialogues do not require manual annotation, as opposed to a Wizard-of-Oz setup BIBREF17, which is a common approach utilized in other datasets BIBREF0. It has been shown that such datasets suffer from substantial annotation errors BIBREF18. Thirdly, using a simulator greatly simplifies the data collection task and instructions as only paraphrasing is needed to achieve a natural dialogue. This is particularly important for creating a large dataset spanning multiple domains. The 20 domains present across the train, dev and test datasets are listed in Table TABREF10, as are the details regarding which domains are present in each of the datasets. We create synthetic implementations of a total of 45 services or APIs over these domains. Our simulator framework interacts with these services to generate dialogue outlines, which are structured representations of dialogue semantics. We then use a crowd-sourcing procedure to paraphrase these outlines to natural language utterances. Our novel crowd-sourcing procedure preserves all annotations obtained from the simulator and does not require any extra annotations after dialogue collection. In this section, we describe these steps briefly and then present analyses of the collected dataset. All the services are implemented using a SQL engine. Since entity attributes are often correlated, we decided not to sample synthetic entities and instead relied on sampling entities from Freebase. The dialogue simulator interacts with the services to generate valid dialogue outlines. The simulator consists of two agents playing the roles of the user and the system. Both agents interact with each other using a finite set of actions specified through dialogue acts over a probabilistic automaton designed to capture varied dialogue trajectories. At the start of the conversation, the user agent is seeded with a scenario, which is a sequence of intents to be fulfilled. The user agent generates dialogue acts to be output and combines them with values retrieved from the service/API to create the user actions. The system agent responds by following a similar procedure but also ensures that the generated flows are valid. We identified over 200 distinct scenarios for the training set each consisting up to 5 intents from various domains. Finally, the dialogue outlines generated are paraphrased into a natural conversation by crowd workers. We ensure that the annotations for the dialogue state and slots generated by the simulator are preserved and hence need no other annotation. We omit details for brevity: please refer to BIBREF19 for more details. The entire dataset consists of over 16K dialogues spanning multiple domains. Overall statistics of the dataset and comparison with other datasets can be seen in Table TABREF9. Figure FIGREF8 shows the details of the distribution of dialogue lengths across single-domain and multi-domain dialogues. The single-domain dialogues in our dataset contain an average of 15.3 turns, whereas the multi-domain ones contain 23 turns on average. Figure FIGREF8 shows the frequency of the different dialogue acts contained in the dataset. The dataset also contains a significant number of unseen domains/APIs in the dev and test sets. 77% of the dialogue turns in the test set and 45% of the turns in dev set contain at least one service not present in the training set. This facilitates the development of models which can generalize to new domains with very few labelled examples. ### Submissions
The submissions from 25 teams included a variety of approaches and innovative solutions to specific problems posed by this dataset. For the workshop, we received submissions from 9 of these teams. In this section, we provide a short summary of the approaches followed by these teams. For effective generalization to unseen APIs, most teams used pre-trained encoders to encode schema element descriptions. Unless otherwise mentioned, a pre-trained BERT BIBREF20 encoder was used. Team 2 BIBREF21: This was the only paper not using a pre-trained encoder, thus providing another important baseline. They rely on separate RNNs to encode service, slot and intent descriptions, and a BiRNN to encode dialogue history. Slot values are inferred using a TRADE-like encoder-decoder setup with a 3-way slot status gate, using the utterance encoding and schema element embeddings as context. Team 5 BIBREF22: They predict values for categorical slots using a softmax over all candidate values. Non-categorical slot values are predicted by first predicting the status of each slot and then using a BiLSTM-CRF layer for BIO tagging BIBREF23. They also utilize a slot adoption tracker to predict if the values proposed by the system are accepted by the user. Team 9 BIBREF24: This team submitted the winning entry, beating the second-placed team by around 9% in terms of joint goal accuracy. They use two separate models for categorical and non-categorical slots, and treat numerical categorical slots as non-categorical. They also use the entire dialogue history as input. They perform data augmentation by back translation between English and Chinese, which seems to be one of the distinguishing factors resulting in a much higher accuracy. Team 12 BIBREF25: They use auxiliary binary features to connect previous intent to current intent, slots to dialogue history and source slots to target slots for slot transfer. Non-categorical slots are modeled similar to question answering by adding a null token and predicting spans for slot values. In-domain and cross-domain slot transfers are modeled as separate binary decisions by passing the slot descriptions as additional inputs. Team 16 BIBREF26: They convert the tracking task for both categorical and non-categorical slots into a question answering task by feeding in the schema and the previous turns as the context. Similar to the baseline model, prediction is performed in two stages. The status of each slot (active/inactive/dontcare) is predicted using a classifier, following which the value is predicted as a span in the context. The same network is used for the different prediction tasks but the leading token and separator tokens used are different. They observe large gains by fine-tuning the schema embeddings and increasing the number of past turns fed as context. Team 23 BIBREF27: They use a large scale multi-task model utilizing a single pass of a BERT based model for all tasks. Embeddings are calculated for the intents and slot value by using dialogue history, service and slot descriptions, possible values for categorical slots and are used for the various predictions. Anonymous Team A BIBREF28: We could not identify which team submitted this model. They use multi-head attention twice to obtain domain-conditioned and slot-conditioned representations of the dialogue history. These representations are concatenated to obtain the full context which is used for the various predictions. Anonymous Team B BIBREF29: We could not identify which team submitted this model. They use separate NLU systems for the sub tasks of predicting intents, requested slots, slot status, categorical and non-categorical slot values. They use a rule-based DST system with a few additions resulting in significant improvement. The improvements include adding dropout to intent prediction to account for train-test mismatch, using the entire predicted slot status distribution and separate binary predictions for slot transfer. Anonymous Team C BIBREF30: They use a two-stage model with a candidate tracker for NLU and a candidate classifier to update the dialogue state. A slot tagger identifies slot values, which are used to update the candidate tracker. The candidate classifier uses the utterances and slot/intent descriptions to predict the final dialogue state. They also use an additional loss to penalize incorrect prediction on which slots appear in the current turn. ### Evaluation
We consider the following metrics for automatic evaluation of different submissions. Joint goal accuracy has been used as the primary metric to rank the submissions. Active Intent Accuracy: The fraction of user turns for which the active intent has been correctly predicted. Requested Slot F1: The macro-averaged F1 score for requested slots over all eligible turns. Turns with no requested slots in ground truth and predictions are skipped. Average Goal Accuracy: For each turn, we predict a single value for each slot present in the dialogue state. This is the average accuracy of predicting the value of a slot correctly. Joint Goal Accuracy: This is the average accuracy of predicting all slot assignments for a given service in a turn correctly. In order to better reflect model performance in our task's specific setting, we introduce changes in the definitions of evaluation metrics from prior work. These are listed below: [leftmargin=*] Joint goal accuracy calculation: Traditionally, joint goal accuracy has been defined as the accuracy of predicting the dialogue state for all domains correctly. This is not practical in our setup, as the large number of services would result in near zero joint goal accuracy if the traditional definition is used. Furthermore, an incorrect dialogue state prediction for a service in the beginning of a dialogue degrades the joint goal accuracy for all future turns, even if the predictions for all other services are correct. Hence, joint goal accuracy calculated this way may not provide as much insight into the performance on different services. To address these concerns, only the services which are active or pertinent in a turn are included in the dialogue state. Thus, a service ceases to be a part of the dialogue state once its intent has been fulfilled. Fuzzy matching for non-categorical slot values: The presence of non-categorical slots is another distinguishing feature of our dataset. These slots don't have a predefined vocabulary, and their values are predicted as a substring or span of the past user or system utterances. Drawing inspiration from the metrics used for slot tagging in spoken language understanding, we use a fuzzy matching score for non-categorical slots to reward partial matches with the ground truth. Average goal accuracy: To calculate average goal accuracy, we do not take into account instances when both the ground truth and the predicted values for a slot are empty. Since for a given slot, a large number of utterances have an empty assignment, models can achieve a relatively high average goal accuracy just by predicting an empty assignment for each slot unless specifically excluded as in our evaluation. ### Results
The test set contains a total of 21 services, among which 6 services are also present in the training set (seen services), whereas the remaining 15 are not present in the training set (unseen services). Table TABREF11 shows the evaluation metrics for the different submissions obtained on the test set. It also lists the performance of different submissions on seen and unseen services, helping evaluate the effectiveness in zero-shot settings. Team 9 achieved a very high joint goal accuracy of 86.53%, around 9% higher than the second-placed team. We observed the following trends across submissions: For unseen services, performance on categorical slots is comparable to that on non-categorical slots. On the other hand, for seen services, the performance on categorical slots is better. This could be because there is less signal to differentiate between the different possible values for a categorical slot when they have not been observed in the training set. The winning team's performance on seen services is similar to that of the other top teams. However, the winning team has a considerable edge on unseen services, outperforming the second team by around 12% in terms of joint goal accuracy. This margin was observed across both categorical and non-categorical slots. Among unseen services, when looking at services belonging to unseen domains, the winning team was ahead of the other teams by at least 15%. The performance on categorical slots for unseen domains was about the same as that for seen services and domains. For other teams, there was at least a 20% drop in accuracy of categorical slots in unseen domains vs seen domains and services. The joint goal accuracy of most of the models was worse by 15 percentage points on an average on the test set as compared to the dev set. This could be because the test set contains a much higher proportion of turns with at least one unseen services as compared to the dev set (77% and 45% respectively). ### Summary
In this paper, we summarized the Schema-Guided Dialogue State Tracking task conducted at the Eighth Dialogue System Technology Challenge. This task challenged participants to develop dialogue state tracking models for large scale virtual assistants, with particular emphasis on joint modeling across different domains and APIs for data-efficiency and zero-shot generalization to new/unseen APIs. In order to encourage the development of such models, we constructed a new dataset spanning 16 domains (and 4 new domains in dev and test sets), defining multiple APIs with overlapping functionality for each of these domains. We advocated the use of schema-guided approach to building large-scale assistants, facilitating data-efficient joint modeling across domains while reducing maintenance workload. The Schema-Guided Dialogue dataset released as part of this task is the first to highlight many of the aforementioned challenges. As a result, this task led to the development of several models utilizing the schema-guided approach for dialogue state tracking. The models extensively utilized pre-trained encoders like BERT BIBREF20, XLNet BIBREF31 etc. and employed data augmentation techniques to achieve effective zero-shot generalization to new APIs. The proposed schema-guided approach is fairly general and can be used to develop other dialogue system components such as language understanding, policy and response generation. We plan to explore them in future works. ### Summary ::: Acknowledgements
The authors thank Guan-Lin Chao, Amir Fayazi and Maria Wang for their advice and assistance. Figure 1: Example schema for a digital wallet service. Figure 2: Dialogue state tracking labels after each user utterance in a dialogue in the context of two different flight services. Under the schema-guided approach, the annotations are conditioned on the schema (extreme left/right) of the underlying service. Figure 3: Detailed statistics of the SGD dataset. Table 1: Comparison of our SGD dataset to existing related datasets for task-oriented dialogue. Note that the numbers reported are for the training portions for all datasets except FRAMES, where the numbers for the complete dataset are reported. Table 2: The total number of intents (services in parentheses) and dialogues for each domain across train1, dev2 and test3 sets. Superscript indicates the datasets in which dialogues from the domain are present. Multi-domain dialogues contribute to counts of each domain. The domain Services includes salons, dentists, doctors, etc. Table 3: The best submission from each team, ordered by the joint goal accuracy on the test set. Teams marked with * submitted their papers to the workshop. We could not identify the teams for three of the submitted papers. | back translation between English and Chinese |
What is the purpose of the battle scene from the story?
A. To accurately depict a significant battle from the Crusades
B. To associate tobacco products with masculinity, brotherhood, and pride
C. To illustrate the powerful bonds of allegiance among soldiers on the battlefield
D. To reveal how the King Phillip's cowardice initiated the downfall of one of the world's greatest armies
| ... After a Few Words ... by Seaton McKettrig Illustrated by Summer [Transcriber's Note: This etext was produced from Analog October 1962. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed.] This is a science-fiction story. History is a science; the other part is, as all Americans know, the most fictional field we have today. He settled himself comfortably in his seat, and carefully put the helmet on, pulling it down firmly until it was properly seated. For a moment, he could see nothing. Then his hand moved up and, with a flick of the wrist, lifted the visor. Ahead of him, in serried array, with lances erect and pennons flying, was the forward part of the column. Far ahead, he knew, were the Knights Templars, who had taken the advance. Behind the Templars rode the mailed knights of Brittany and Anjou. These were followed by King Guy of Jerusalem and the host of Poitou. He himself, Sir Robert de Bouain, was riding with the Norman and English troops, just behind the men of Poitou. Sir Robert turned slightly in his saddle. To his right, he could see the brilliant red-and-gold banner of the lion-hearted Richard of England— gules, in pale three lions passant guardant or . Behind the standard-bearer, his great war horse moving with a steady, measured pace, his coronet of gold on his steel helm gleaming in the glaring desert sun, the lions of England on his firm-held shield, was the King himself. Further behind, the Knights Hospitallers protected the rear, guarding the column of the hosts of Christendom from harassment by the Bedouins. "By our Lady!" came a voice from his left. "Three days out from Acre, and the accursed Saracens still elude us." Sir Robert de Bouain twisted again in his saddle to look at the knight riding alongside him. Sir Gaeton de l'Arc-Tombé sat tall and straight in his saddle, his visor up, his blue eyes narrowed against the glare of the sun. Sir Robert's lips formed a smile. "They are not far off, Sir Gaeton. They have been following us. As we march parallel to the seacoast, so they have been marching with us in those hills to the east." "Like the jackals they are," said Sir Gaeton. "They assail us from the rear, and they set up traps in our path ahead. Our spies tell us that the Turks lie ahead of us in countless numbers. And yet, they fear to face us in open battle." "Is it fear, or are they merely gathering their forces?" "Both," said Sir Gaeton flatly. "They fear us, else they would not dally to amass so fearsome a force. If, as our informers tell us, there are uncounted Turks to the fore, and if, as we are aware, our rear is being dogged by the Bedouin and the black horsemen of Egypt, it would seem that Saladin has at hand more than enough to overcome us, were they all truly Christian knights." "Give them time. We must wait for their attack, sir knight. It were foolhardy to attempt to seek them in their own hills, and yet they must stop us. They will attack before we reach Jerusalem, fear not." "We of Gascony fear no heathen Musselman," Sir Gaeton growled. "It's this Hellish heat that is driving me mad." He pointed toward the eastern hills. "The sun is yet low, and already the heat is unbearable." Sir Robert heard his own laugh echo hollowly within his helmet. "Perhaps 'twere better to be mad when the assault comes. Madmen fight better than men of cooler blood." He knew that the others were baking inside their heavy armor, although he himself was not too uncomfortable. Sir Gaeton looked at him with a smile that held both irony and respect. "In truth, sir knight, it is apparent that you fear neither men nor heat. Nor is your own blood too cool. True, I ride with your Normans and your English and your King Richard of the Lion's Heart, but I am a Gascon, and have sworn no fealty to him. But to side with the Duke of Burgundy against King Richard—" He gave a short, barking laugh. "I fear no man," he went on, "but if I had to fear one, it would be Richard of England." Sir Robert's voice came like a sword: steely, flat, cold, and sharp. "My lord the King spoke in haste. He has reason to be bitter against Philip of France, as do we all. Philip has deserted the field. He has returned to France in haste, leaving the rest of us to fight the Saracen for the Holy Land leaving only the contingent of his vassal the Duke of Burgundy to remain with us." "Richard of England has never been on the best of terms with Philip Augustus," said Sir Gaeton. "No, and with good cause. But he allowed his anger against Philip to color his judgment when he spoke harshly against the Duke of Burgundy. The Duke is no coward, and Richard Plantagenet well knows it. As I said, he spoke in haste." "And you intervened," said Sir Gaeton. "It was my duty." Sir Robert's voice was stubborn. "Could we have permitted a quarrel to develop between the two finest knights and warleaders in Christendom at this crucial point? The desertion of Philip of France has cost us dearly. Could we permit the desertion of Burgundy, too?" "You did what must be done in honor," the Gascon conceded, "but you have not gained the love of Richard by doing so." Sir Robert felt his jaw set firmly. "My king knows I am loyal." Sir Gaeton said nothing more, but there was a look in his eyes that showed that he felt that Richard of England might even doubt the loyalty of Sir Robert de Bouain. Sir Robert rode on in silence, feeling the movement of the horse beneath him. There was a sudden sound to the rear. Like a wash of the tide from the sea came the sound of Saracen war cries and the clash of steel on steel mingled with the sounds of horses in agony and anger. Sir Robert turned his horse to look. The Negro troops of Saladin's Egyptian contingent were thundering down upon the rear! They clashed with the Hospitallers, slamming in like a rain of heavy stones, too close in for the use of bows. There was only the sword against armor, like the sound of a thousand hammers against a thousand anvils. "Stand fast! Stand fast! Hold them off!" It was the voice of King Richard, sounding like a clarion over the din of battle. Sir Robert felt his horse move, as though it were urging him on toward the battle, but his hand held to the reins, keeping the great charger in check. The King had said "Stand fast!" and this was no time to disobey the orders of Richard. The Saracen troops were coming in from the rear, and the Hospitallers were taking the brunt of the charge. They fought like madmen, but they were slowly being forced back. The Master of the Hospitallers rode to the rear, to the King's standard, which hardly moved in the still desert air, now that the column had stopped moving. The voice of the Duke of Burgundy came to Sir Robert's ears. "Stand fast. The King bids you all to stand fast," said the duke, his voice fading as he rode on up the column toward the knights of Poitou and the Knights Templars. The Master of the Hospitallers was speaking in a low, urgent voice to the King: "My lord, we are pressed on by the enemy and in danger of eternal infamy. We are losing our horses, one after the other!" "Good Master," said Richard, "it is you who must sustain their attack. No one can be everywhere at once." The Master of the Hospitallers nodded curtly and charged back into the fray. The King turned to Sir Baldwin de Carreo, who sat ahorse nearby, and pointed toward the eastern hills. "They will come from there, hitting us in the flank; we cannot afford to amass a rearward charge. To do so would be to fall directly into the hands of the Saracen." A voice very close to Sir Robert said: "Richard is right. If we go to the aid of the Hospitallers, we will expose the column to a flank attack." It was Sir Gaeton. "My lord the King," Sir Robert heard his voice say, "is right in all but one thing. If we allow the Egyptians to take us from the rear, there will be no need for Saladin and his Turks to come down on our flank. And the Hospitallers cannot hold for long at this rate. A charge at full gallop would break the Egyptian line and give the Hospitallers breathing time. Are you with me?" "Against the orders of the King?" "The King cannot see everything! There are times when a man must use his own judgment! You said you were afraid of no man. Are you with me?" After a moment's hesitation, Sir Gaeton couched his lance. "I'm with you, sir knight! Live or die, I follow! Strike and strike hard!" "Forward then!" Sir Robert heard himself shouting. "Forward for St. George and for England!" "St. George and England!" the Gascon echoed. Two great war horses began to move ponderously forward toward the battle lines, gaining momentum as they went. Moving in unison, the two knights, their horses now at a fast trot, lowered their lances, picking their Saracen targets with care. Larger and larger loomed the Egyptian cavalrymen as the horses changed pace to a thundering gallop. The Egyptians tried to dodge, as they saw, too late, the approach of the Christian knights. Sir Robert felt the shock against himself and his horse as the steel tip of the long ash lance struck the Saracen horseman in the chest. Out of the corner of his eye, he saw that Sir Gaeton, too, had scored. The Saracen, impaled on Sir Robert's lance, shot from the saddle as he died. His lighter armor had hardly impeded the incoming spear-point, and now his body dragged it down as he dropped toward the desert sand. Another Moslem cavalryman was charging in now, swinging his curved saber, taking advantage of Sir Robert's sagging lance. There was nothing else to do but drop the lance and draw his heavy broadsword. His hand grasped it, and it came singing from its scabbard. The Egyptian's curved sword clanged against Sir Robert's helm, setting his head ringing. In return, the knight's broadsword came about in a sweeping arc, and the Egyptian's horse rode on with the rider's headless body. Behind him, Sir Robert heard further cries of "St. George and England!" The Hospitallers, taking heart at the charge, were going in! Behind them came the Count of Champagne, the Earl of Leister, and the Bishop of Beauvais, who carried a great warhammer in order that he might not break Church Law by shedding blood. Sir Robert's own sword rose and fell, cutting and hacking at the enemy. He himself felt a dreamlike detachment, as though he were watching the battle rather than participating in it. But he could see that the Moslems were falling back before the Christian onslaught. And then, quite suddenly, there seemed to be no foeman to swing at. Breathing heavily, Sir Robert sheathed his broadsword. Beside him, Sir Gaeton did the same, saying: "It will be a few minutes before they can regroup, sir knight. We may have routed them completely." "Aye. But King Richard will not approve of my breaking ranks and disobeying orders. I may win the battle and lose my head in the end." "This is no time to worry about the future," said the Gascon. "Rest for a moment and relax, that you may be the stronger later. Here—have an Old Kings ." He had a pack of cigarettes in his gauntleted hand, which he profferred to Sir Robert. There were three cigarettes protruding from it, one slightly farther than the others. Sir Robert's hand reached out and took that one. "Thanks. When the going gets rough, I really enjoy an Old Kings ." He put one end of the cigarette in his mouth and lit the other from the lighter in Sir Gaeton's hand. "Yes, sir," said Sir Gaeton, after lighting his own cigarette, " Old Kings are the greatest. They give a man real, deep-down smoking pleasure." "There's no doubt about it, Old Kings are a man's cigarette." Sir Robert could feel the soothing smoke in his lungs as he inhaled deeply. "That's great. When I want a cigarette, I don't want just any cigarette." "Nor I," agreed the Gascon. " Old Kings is the only real cigarette when you're doing a real man's work." "That's for sure." Sir Robert watched a smoke ring expand in the air. There was a sudden clash of arms off to their left. Sir Robert dropped his cigarette to the ground. "The trouble is that doing a real he-man's work doesn't always allow you to enjoy the fine, rich tobaccos of Old Kings right down to the very end." "No, but you can always light another later," said the Gascon knight. King Richard, on seeing his army moving suddenly toward the harassed rear, had realized the danger and had charged through the Hospitallers to get into the thick of the fray. Now the Turks were charging down from the hills, hitting—not the flank as he had expected, but the rear! Saladin had expected him to hold fast! Sir Robert and Sir Gaeton spurred their chargers toward the flapping banner of England. The fierce warrior-king of England, his mighty sword in hand, was cutting down Turks as though they were grain-stalks, but still the Saracen horde pressed on. More and more of the terrible Turks came boiling down out of the hills, their glittering scimitars swinging. Sir Robert lost all track of time. There was nothing to do but keep his own great broadsword moving, swinging like some gigantic metronome as he hacked down the Moslem foes. And then, suddenly, he found himself surrounded by the Saracens! He was isolated and alone, cut off from the rest of the Christian forces! He glanced quickly around as he slashed another Saracen from pate to breastbone. Where was Sir Gaeton? Where were the others? Where was the red-and-gold banner of Richard? He caught a glimpse of the fluttering banner far to the rear and started to fall back. And then he saw another knight nearby, a huge man who swung his sparkling blade with power and force. On his steel helm gleamed a golden coronet! Richard! And the great king, in spite of his prowess was outnumbered heavily and would, within seconds, be cut down by the Saracen horde! Without hesitation, Sir Robert plunged his horse toward the surrounded monarch, his great blade cutting a path before him. He saw Richard go down, falling from the saddle of his charger, but by that time his own sword was cutting into the screaming Saracens and they had no time to attempt any further mischief to the King. They had their hands full with Sir Robert de Bouain. He did not know how long he fought there, holding his charger motionless over the inert body of the fallen king, hewing down the screaming enemy, but presently he heard the familiar cry of "For St. George and for England" behind him. The Norman and English troops were charging in, bringing with them the banner of England! And then Richard was on his feet, cleaving the air about him with his own broadsword. Its bright edge, besmeared with Saracen blood, was biting viciously into the foe. The Turks began to fall back. Within seconds, the Christian knights were boiling around the embattled pair, forcing the Turks into retreat. And for the second time, Sir Robert found himself with no one to fight. And then a voice was saying: "You have done well this day, sir knight. Richard Plantagenet will not forget." Sir Robert turned in his saddle to face the smiling king. "My lord king, be assured that I would never forget my loyalty to my sovereign and liege lord. My sword and my life are yours whenever you call." King Richard's gauntleted hand grasped his own. "If it please God, I shall never ask your life. An earldom awaits you when we return to England, sir knight." And then the king mounted his horse and was running full gallop after the retreating Saracens. Robert took off his helmet. He blinked for a second to adjust his eyes to the relative dimness of the studio. After the brightness of the desert that the televicarion helmet had projected into his eyes, the studio seemed strangely cavelike. "How'd you like it, Bob?" asked one of the two producers of the show. Robert Bowen nodded briskly and patted the televike helmet. "It was O.K.," he said. "Good show. A little talky at the beginning, and it needs a better fade-out, but the action scenes were fine. The sponsor ought to like it—for a while, at least." "What do you mean, 'for a while'?" Robert Bowen sighed. "If this thing goes on the air the way it is, he'll lose sales." "Why? Commercial not good enough?" " Too good! Man, I've smoked Old Kings , and, believe me, the real thing never tasted as good as that cigarette did in the commercial!" | B. To associate tobacco products with masculinity, brotherhood, and pride |
How did suspension help the crew?
A. They could survive without oxygen.
B. They could live on an inhospitable planet.
C. They could survive with lack of gravity.
D. They could travel through space for a long distance.
| CAPTAIN CHAOS By D. ALLEN MORRISSEY Science equipped David Corbin with borrowed time; sent him winging out in a state of suspension to future centuries ... to a dark blue world whose only defense was to seal tight the prying minds of foolish interlopers. [Transcriber's Note: This etext was produced from Planet Stories November 1952. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed.] I heard the voice as I opened my eyes. I was lying down, still not aware of where I was, waiting for the voice. "Your name is David Corbin. Do you understand?" I looked in the direction of the sound. Above my feet a bulkhead loomed. There were round dials set in a row above a speaker. Over the mesh-covered speaker, two knobs glowed red. I ran the words over in my sluggish mind, thinking about an answer. The muscles in my throat tightened up in reflex as I tried to bring some unity into the jumble of thoughts and ideas that kept forming. One word formed out of the rush of anxiety. "No." I shouted a protest against the strangeness of the room. I looked to the right, my eyes following the curving ceiling that started at the cot. The curve met another straight bulkhead on the left. I was in a small room, gray in color, like dull metal. Overhead a bright light burned into my vision. I wondered where in the universe I was. "Your name is David Corbin. If you understand, press button A on your right." I stared at the speaker in the wall. The mesh-covered hole and the two lights looked like a caricature of a face, set in a panel of dials. I twisted my head to look for the button. I pushed away from the close wall but I couldn't move. I reached down to the tightness that held my body, found the wide strap that held me and fumbled with the buckle. I threw it off and pushed myself up from the hard cot. I heard myself yell in surprise as I floated up towards the light overhead. I was weightless. How do you describe being weightless when you are born into a world bound by gravity. I twisted and shut my eyes in terror. There was no sensation of place, no feeling of up or down, no direction. My back bumped against the ceiling and I opened my eyes to stare at the cot and floor. I was concentrating too hard on remembering to be frightened for long. I pushed away from the warm metal and the floor moved up to meet me. "If you understand, press button A on your right." What should I understand? That I was floating in a room that had a curved wall ... that nothing was right in this hostile room? When I reached the cot I held it and drew myself down. I glanced at the planes of the room, trying to place it with other rooms I could see in my mind. Gray walls with a crazy curved ceiling ... a door to my left that appeared to be air tight. I stared at my familiar hands. I rubbed them across my face, feeling the solidity of flesh and bone, afraid to think too hard about myself. "My name ... my name is...." "Your name is David Corbin." I stared at the speaker. How long did this go on? The name meant nothing to me, but I thought about it, watching the relentless lights that shone below the dials. I stood up slowly and looked at myself. I was naked except for heavy shorts, and there was no clue to my name in the pockets. The room was warm and the air I had been breathing was good but it seemed wrong to be dressed like this. I didn't know why. I thought about insanity, and the room seemed to fit my thoughts. When the voice repeated the message again I had to act. Walking was like treading water that couldn't be seen or felt. I floated against the door, twisting the handle in fear that it wouldn't turn. The handle clanged as I pushed it down and I stared at the opposite wall of a narrow gray passageway. I pushed out into it and grasped the metal rail that ran along the wall. I reasoned it was there to propel yourself through the passageway in this weightless atmosphere. It was effortless to move. I turned on my side like a swimmer and went hand over hand, shooting down the corridor. I braced against forward motion and stopped against a door at the end. Behind me I could see the opened door I had left, and the thought of that questioning voice made me want to move. I swung the door open, catching a glimpse of a room crowded with equipment and.... I will always remember the scream of terror, the paralyzing fright of what I saw through the portholes in the wall of the room. I saw the blackest night, pierced by brilliance that blinded me. There was no depth to the searing brightness of countless stars. They seemed to press against the glass, blobs of fire against a black curtain burning into my eyes and brain. It was space. I looked out at deep space, star systems in clusters. I shut my eyes. When I looked again I knew where I was. Why the little room had been shaped like quarter round. Why I drifted weightlessly. Why I was.... David Corbin. I knew more of the puzzle. Something was wrong. After the first shock of looking out, I accepted the fact that I was in a space ship, yet I couldn't read the maps that were fastened to a table, nor understand the function or design of the compact machinery. WHY, Why, Why? The thought kept pounding at me. I was afraid to touch anything in the room. I pressed against the clear window, wondering if the stars were familiar. I had a brief vivid picture of a night sky on Earth. This was not the same sky. Back in the room where I had awakened, I touched the panel with the glowing eyes. It had asked me if I understood. Now it must tell me why I didn't. It had to help me, that flat metallic voice that repeated the same words. It must tell me.... "Your name is David Corbin. If you understand, press button A on your right." I pressed the button by the cot. The red lights blinked out as I stood in patient attention, trying to outguess the voice. I recalled a phrase ... some words about precaution. Precaution against forgetting. It was crazy, but I trusted the panel. It was the only thing I saw that could help me, guard me against another shock like seeing outside of the clear portholes. "It is assumed the experiment is a success," the voice said. What experiment? "You have been removed from suspension. Assume manual control of this ship." Control of a ship? Going where? "Do not begin operations until the others are removed from suspension." What others? Tell me what to do. "Rely on instructions for factoring when you check the coordinates. Your maximum deviation from schedule cannot exceed two degrees. Adopt emergency procedures as you see fit. Good luck." The voice snapped off and I laughed hysterically. None of it had made sense, and I cursed whatever madness had put me here. "Tell me what to do," I shouted wildly. I hammered the hard metal until the pain in my hands made me stop. "I can't remember what to do." I held my bruised hands to my mouth, and I knew that was all the message there was. In blind panic I pushed away from the panel. Something tripped me and I fell back in a graceless arc. I pushed away from the floor, barely feeling the pain in my leg, and went into the hall. Pain burned along my leg but I couldn't stop. In the first panic of waking up in strangeness I had missed the other doors in the passage. The first swung back to reveal a deep closet holding five bulky suits. The second room was like my own. A dark haired, deep chested man lay on the cot. His muscular body was secured by a wide belt. He was as still as death, motionless without warmth or breath as I hovered over him. I couldn't remember his face. The next room held another man. He was young and wiry, like an athlete cast in marble, dark haired and big jawed. A glassy eye stared up when I rolled back his eyelid. The eyelid remained open until I closed it and went on. Another room ... another man ... another stranger. This man was tall and raw boned, light of skin and hair, as dead as the others. A flat, illogical voice had instructed me to revive these men. I shivered in spite of the warmth of the room, studying the black box that squatted on a shelf by his head. My hand shook when I touched the metal. I dared not try to operate anything. Revive the others ... instructions without knowledge were useless to me. I stopped looking into the doors in the passageway and went back to the room with the portholes. Everything lay in readiness, fastened down star charts, instruments, glittering equipment. There was no feeling of disorder or use in the room. It waited for human hands to make it operate. Not mine. Not now. I went past the room into another, where the curves were more sharp. I could visualize the tapering hull leading to the nose of the ship. This room was filled with equipment that formed a room out of the bordered area I stood in. I sat in the deep chair facing the panel of dials and instruments, in easy reach. I ran my hands over the dials, the rows of smooth colored buttons, wondering. The ports on the side were shielded and I stared out at static energy, hung motionless in a world of searing light. There was no distortion, no movement outside and I glanced back at the dials. What speeds were they recording? What speeds and perhaps, what distance? It was useless to translate the markings. They stood for anything I might guess, and something kept pricking my mind, telling me I had no time to guess. I thought of time again. I was supposed to act according to ... plan. Did that mean ... in time ... in time. I went back down the passageway. The fourth small room was the same. Except for the woman. She lay on a cot, young and beautiful, even in the death-like immobility I had come to accept. Her beauty was graceful lines of face and her figure—smooth tapering legs, soft curves that were carved out of flesh colored stone. Yet not stone. I held her small hand, then put it back on the cot. Her attire was brief like the rest of us, shorts and a man's shirt. Golden hair curled up around her lovely face. I wondered if she would ever smile or move that graceful head. I rolled back her eyelid and looked at a deep blue eye that stared back in glassy surprise. Four people in all, depending on a blind helpless fool who didn't know their names or the reason for that dependence. I sat beside her on the cot until I could stand it no longer. Searching the ship made me forget my fear. I hoped I would find some answers. I went from the nose to the last bulkhead in a frenzy of floating motion, looking behind each door until I went as far as I could. There were two levels to the ship. They both ended in the lead shield that was set where the swell of the curve was biggest. It meant the engine or engines took up half the ship, cut off from the forward half by the instrument studded shield. I retraced my steps and took a rough estimate of size. The ship, as I called it, was at least four hundred feet long, fifty feet in diameter on the inside. The silence was a force in itself, pressing down from the metal walls, driving me back to the comforting smallness of the room where I had been reborn. I laughed bitterly, thinking about the aptness of that. I had literally been reborn in this room, equipped with half ideas, and no point to start from, no premise to seek. I sensed the place to start from was back in the room. I searched it carefully. Minutes later I realized the apparatus by the cot was different. It was the same type of black box, but out from it was a metal arm, bent in a funny angle. At the tip of the arm, a needle gleamed dully and I rubbed the deep gash on my leg. I bent the arm back until the angle looked right. It was then I realized the needle came to a spot where it could have hit my neck when I lay down. My shout of excitement rang out in the room, as I pictured the action of the extended arm. I lost my sudden elation in the cabin where the girl lay. The box behind her head was completely closed, and it didn't yield to the pressure I applied. It had a cover, but no other opening where an arm could extend. I ran my fingers over the unbroken surface, prying over the thin crack at the base helplessly. If some sort of antidote was to be administered manually I was lost. I had no knowledge of what to inject or where to look for it. The chamber of the needle that had awakened me was empty. That meant a measured amount. In the laboratory on the lower level I went over the rows of cans and tubes fastened to the shelves. There were earths and minerals, seeds and chemicals, testing equipment in compact drawers, but nothing marked for me. I wondered if I was an engineer or a pilot, or perhaps a doctor sent along to safeguard the others. Complete amnesia would have been terrible enough but this half knowledge, part awareness and association with the ship was a frightening force that seemed ready to break out of me. I went back to the cabin where the powerful man lay. I had to risk failure with one of them. I didn't want it to be the girl. I fought down the thought that he might be the key man, remembering the voice that had given the message. It was up to me, and soon. The metal in the box would have withstood a bullet. It couldn't be pried apart, and I searched again and again for a release mechanism. I found it. I swung the massive cover off and set it down. The equipment waited for the touch of a button and it went into operation. I stepped back as the tubes glowed to life and the arm swung down with the gleaming needle. The needle went into the corded neck of the man. The fluid chamber drained under pressure and the arm moved back. I stood by the man for long minutes. Finally it came. He stirred restlessly, closing his hands into fists. The deep chest rose and fell unevenly as he breathed. Finally the eyes opened and he looked at me. I watched him adjust to the room. It was in his eyes, wide at first, moving about the confines of the room back to me. "It looks like we made it," he said. "Yes." He unfastened the belt and sat up. I pushed him back as he floated up finding little humor in the comic expression on his face. "No gravity," he grunted and sat back. "You get used to it fast," I answered. I thought of what to say as he watched me. "How do you feel?" He shrugged at the question. "Fine, I guess. Funny, I can't remember." He saw it in my face, making him stop. "I can't remember dropping off to sleep," he finished. I held his hard arm. "What else? How much do you remember?" "I'm all right," he answered. "There aren't supposed to be any effects from this." "Who is in charge of this ship?" I asked. He tensed suddenly. "You are, sir. Why?" I moved away from the cot. "Listen, I can't remember. I don't know your name or anything about this ship." "What do you mean? What can't you remember?" he asked. He stood up slowly, edging around towards the door. I didn't want to fight him. I wanted him to understand. "Look, I'm in trouble. Nothing fits, except my name." "You don't know me?" "No." "Are you serious?" "Yes, yes. I don't know why but it's happened." He let his breath out in a whistle. "For God's sake. Any bump on your head?" "I feel all right physically. I just can't place enough." "The others. What about the others?" he blurted. "I don't know. You're the first besides myself. I don't know how I stumbled on the way to revive you." He shook his head, watching me like I was a freak. "Let's check the rest right away." "Yes. I've got to know if they are like me. I'm afraid to think they might be." "Maybe it's temporary. We can figure something out." II The second man, the dark haired one, opened his eyes and recognized us. He asked questions in rapid fire excitement. The third man, the tall Viking, was all right until he moved. The weightless sensation made him violently sick. We put him back on the cot, securing him again with the belt, but the sight of us floating made him shake. He was retching without results when we drifted out. I followed him to the girl's quarters. "What about her. Why is she here?" I asked my companion. He lifted the cover from the apparatus. "She's the chemist in the crew." "A girl?" "Dr. Thiesen is an expert, trained for this," he said. I looked at her. She looked anything but like a chemist. "There must be men who could have been sent. I've been wondering why a girl." "I don't know why, Captain. You tried to stop her before. Age and experience were all that mattered to the brass." "It's a bad thing to do." "I suppose. The mission stated one chemist." "What is the mission of this ship?" I asked. He held up his hand. "We'd better wait, sir. Everything was supposed to be all right on this end. First you, then Carl, sick to his stomach." "Okay. I'll hold the questions until we see about her." We were out of luck with the girl. She woke up and she was frightened. We questioned her and she was coherent but she couldn't remember. I tried to smile as I sat on the cot, wondering what she was thinking. "How do you feel?" I asked. Her face was a mask of wide-eyed fear as she shook her head. "Can you remember?" "I don't know." Blue eyes stared at me in fear. Her voice was low. "Do you know my name?" The question frightened her. "Should I? I feel so strange. Give me a minute to think." I let her sit up slowly. "Do you know your name?" She tightened up in my arms. "Yes. It's...." She looked at us for help, frightened by the lack of clothing we wore, by the bleak room. Her eyes circled the room. "I'm afraid," she cried. I held her and she shook uncontrollably. "What's happened to me?" she asked. The dark haired man came into the room, silent and watchful. My companion motioned to him. "Get Carl and meet us in Control." The man looked at me and I nodded. "We'll be there in a moment. I'm afraid we've got trouble." He nodded and pushed away from us. The girl screamed and covered her face with her hands. I turned to the other man. "What's your name?" "Croft. John Croft." "John, what are your duties if any?" "Automatic control. I helped to install it." "Can you run this ship? How about the other two?" He hit his hands together. "You fly it, sir. Can't you think?" "I'm trying. I know the ship is familiar, but I've looked it over. Maybe I'm trying too hard." "You flew her from earth until we went into suspension," he said. "I can't remember when," I said. I held the trembling girl against me, shaking my head. He glanced at the girl. "If the calculations are right it was more than a hundred years ago." We assembled in the control room for a council. We were all a little better for being together. John Croft named the others for me. I searched each face without recognition. The blond man was Carl Herrick, a metallurgist. His lean face was white from his spell but he was better. Paul Sample was a biologist, John said. He was lithe and restless, with dark eyes that studied the rest of us. I looked at the girl. She was staring out of the ports, her hands pressed against the transparent break in the smooth wall. Karen Thiesen was a chemist, now frightened and trying to remember. I wasn't in much better condition. "Look, if it comes too fast for me, for any of us, we'll stop. John, you can lead off." "You ask the questions," he said. I indicated the ship. "Where in creation are we going?" "We set out from Earth for a single star in the direction of the center of our Galaxy." "From Earth? How could we?" "Let's move slowly, sir," he said. "We're moving fast. I don't know if you can picture it, but we're going about one hundred thousand miles an hour." "Through space?" "Yes." "What direction?" Paul cut in. "It's a G type star, like our own sun in mass and luminosity. We hope to find a planetary system capable of supporting life." "I can't grasp it. How can we go very far in a lifetime?" "It can be done in two lifetimes," John said quietly. "You said I had flown this ship. You meant before this suspension." "Yes. That's why we can cross space to a near star." "How long ago was it?" "It was set at about a hundred years, sir. Doesn't that fit at all?" "I can't believe it's possible." Carl caught my eye. "Captain, we save this time without aging at all. It puts us near a calculated destination." "We've lost our lifetime." It was Karen. She had been crying silently while we talked. "Don't think about it," Paul said. "We can still pull this out all right if you don't lose your nerve." "What are we to do?" she asked. John answered for me. "First we've got to find out where we are. I know this ship but I can't fly it." "Can I?" I asked. We set up a temporary plan of action. Paul took Karen to the laboratory in an effort to help her remember her job. Carl went back to divide the rations. I was to study the charts and manuals. It was better than doing nothing, and I went into the navigation room and sat down. Earth was an infinitesimal point somewhere behind us on the galactic plane, and no one else was trained to navigate. The ship thundered to life as I sat there. The blast roared once ... twice, then settled into a muted crescendo of sound that hummed through the walls. I went into the control room and watched John at the panel. "I wish I knew what you were doing," I said savagely. "Give it time." "We can't spare any, can we?" I asked. "I wish we knew. What about her—Dr. Thiesen?" "She's in the lab. I don't think that will do much good. She's got to be shocked out of a mental state like that." "I guess you're right," he said slowly. "She's trained to administer the suspension on the return trip." I let my breath out slowly. "I didn't think about that." "We couldn't even get part way back in a lifetime," he said. "How old are you, John?" "Twenty-eight." "What about me?" "Thirty." He stared at the panel in thought for a minutes. "What about shock treatment? It sounds risky." "I know. It's the only thing I could think of. Why didn't everyone react the same?" "That had me wondering for a while. I don't know. Anyway how could you go about making her remember?" "Throw a crisis, some situation at her, I guess." He shrugged, letting his sure hands rest on the panel of dials. I headed back towards the lab. If I could help her I might help myself. I was past the rooms when the horn blasted through the corridor. I turned automatically with the sound, pushing against the rail, towards the control room. Deep in my mind I could see danger, and without questioning why I knew I had to be at Control when the sound knifed through the stillness. John was shouting as I thrust my way into the room. "Turn the ship. There's something dead ahead." I had a glimpse of his contorted face as I dove at the control board. My hands hit buttons, thumbed a switch and then a sudden force threw me to the right. I slammed into the panel on the right, as the pressure of the change dimmed my vision. Reflex made me look up at the radar control screen. It wasn't operating. John let go of the padded chair, grinning weakly. I was busy for a few seconds, feeding compensation into the gyros. Relief flooded through me like warm liquid. I hung on the intercom for support, drawing air into my heaving lungs. "What—made you—think of that," I asked weakly. "Shock treatment." "I must have acted on instinct." "You did. Even for a sick man that was pretty fast," he laughed. "I can think again, John. I know who I am," I shouted. I threw my arms around his massive shoulders. "You did it." "You gave me the idea, Mister, talking about Dr. Thiesen." "It worked. I'm okay," I said in giddy relief. "I don't have to tell you I was scared as hell. I wish you could have seen your face, the look in your eyes when I woke up." "I wouldn't want to wake up like that again." "You're all right now?" he asked. I grinned and nodded an answer. I saw John as he was at the base, big and competent, sweating in the blazing sun. I thought about the rest of the crew too. "We're heading right for a star...." "It's been dead ahead for hours," he grunted. I leaned over and threw the intercom to open. "This is control. Listen ... everyone. I'm over it. Disregard the warning siren ... we were testing the ship." The lab light blinked on as Paul cut in. "What was it ... hey, you said you're all right." "John did it. He hit the alarm figuring I would react. Listen, Paul. Is any one hurt?" "No. Carl is here too. His stomach flopped again but he's okay. What about food. We're supposed to be checked before we eat." "We'll have to go ahead without it. Any change?" "No, I put her to bed. Shall I bring food?" I glanced at John. He rubbed his stomach. "Yes," I answered. "Bring it when you can. I've got to find out where we are." We had to get off course before we ran into the yellow-white star that had been picked for us. Food was set down by me, grew cold and was carried away and I was still rechecking the figures. We were on a line ten degrees above the galactic plane. The parallactic baseline from Earth to the single star could be in error several degrees, or we could be right on the calculated position of the star. The radar confirmed my findings ... and my worst fears. When we set it for direction and distance, the screen glowed to life and recorded the star dead ahead. In all the distant star clusters, only this G type star was thought to have a planetary system like our own. We were out on a gamble to find a planet capable of supporting life. The idea had intrigued scientists before I had first looked up at the night sky. When I was sure the electronically recorded course was accurate for time, I checked direction and speed from the readings and plotted our position. If I was right we were much closer than we wanted to be. The bright pips on the screen gave us the distance and size of the star while we fed the figures into the calculator for our rate of approach. Spectroscopic tests were run on the sun and checked against the figures that had been calculated on Earth. We analyzed temperature, magnetic fields, radial motion, density and luminosity, checking against the standards the scientists had constructed. It was a G type star like our own. It had more density and temperature and suitable planets or not, we had to change course in a hurry. Carl analyzed the findings while we came to a decision. Somewhere along an orbit that might be two hundred miles across, our hypothetical planet circled this star. That distance was selected when the planets in Earth's solar system had proved to be barren. If the observations on this star were correct, we could expect to find a planet in a state of fertility ... if it existed ... if it were suitable for colonization ... if we could find it. | D. They could travel through space for a long distance. |
What did Kolin think about becoming a tree himself?
A. He wanted to be an animal, not a plant.
B. He was intrigued but wanted to try something slightly different.
C. He figured it was an effective way to escape his crew.
D. He refused to give up his own body.
| By H. B. Fyfe THE TALKATIVE TREE Dang vines! Beats all how some plants have no manners—but what do you expect, when they used to be men! All things considered—the obscure star, the undetermined damage to the stellar drive and the way the small planet's murky atmosphere defied precision scanners—the pilot made a reasonably good landing. Despite sour feelings for the space service of Haurtoz, steward Peter Kolin had to admit that casualties might have been far worse. Chief Steward Slichow led his little command, less two third-class ration keepers thought to have been trapped in the lower hold, to a point two hundred meters from the steaming hull of the Peace State . He lined them up as if on parade. Kolin made himself inconspicuous. "Since the crew will be on emergency watches repairing the damage," announced the Chief in clipped, aggressive tones, "I have volunteered my section for preliminary scouting, as is suitable. It may be useful to discover temporary sources in this area of natural foods." Volunteered HIS section! thought Kolin rebelliously. Like the Supreme Director of Haurtoz! Being conscripted into this idiotic space fleet that never fights is bad enough without a tin god on jets like Slichow! Prudently, he did not express this resentment overtly. His well-schooled features revealed no trace of the idea—or of any other idea. The Planetary State of Haurtoz had been organized some fifteen light-years from old Earth, but many of the home world's less kindly techniques had been employed. Lack of complete loyalty to the state was likely to result in a siege of treatment that left the subject suitably "re-personalized." Kolin had heard of instances wherein mere unenthusiastic posture had betrayed intentions to harbor treasonable thoughts. "You will scout in five details of three persons each," Chief Slichow said. "Every hour, each detail will send one person in to report, and he will be replaced by one of the five I shall keep here to issue rations." Kolin permitted himself to wonder when anyone might get some rest, but assumed a mildly willing look. (Too eager an attitude could arouse suspicion of disguising an improper viewpoint.) The maintenance of a proper viewpoint was a necessity if the Planetary State were to survive the hostile plots of Earth and the latter's decadent colonies. That, at least, was the official line. Kolin found himself in a group with Jak Ammet, a third cook, and Eva Yrtok, powdered foods storekeeper. Since the crew would be eating packaged rations during repairs, Yrtok could be spared to command a scout detail. Each scout was issued a rocket pistol and a plastic water tube. Chief Slichow emphasized that the keepers of rations could hardly, in an emergency, give even the appearance of favoring themselves in regard to food. They would go without. Kolin maintained a standard expression as the Chief's sharp stare measured them. Yrtok, a dark, lean-faced girl, led the way with a quiet monosyllable. She carried the small radio they would be permitted to use for messages of utmost urgency. Ammet followed, and Kolin brought up the rear. To reach their assigned sector, they had to climb a forbidding ridge of rock within half a kilometer. Only a sparse creeper grew along their way, its elongated leaves shimmering with bronze-green reflections against a stony surface; but when they topped the ridge a thick forest was in sight. Yrtok and Ammet paused momentarily before descending. Kolin shared their sense of isolation. They would be out of sight of authority and responsible for their own actions. It was a strange sensation. They marched down into the valley at a brisk pace, becoming more aware of the clouds and atmospheric haze. Distant objects seemed blurred by the mist, taking on a somber, brooding grayness. For all Kolin could tell, he and the others were isolated in a world bounded by the rocky ridge behind them and a semi-circle of damp trees and bushes several hundred meters away. He suspected that the hills rising mistily ahead were part of a continuous slope, but could not be sure. Yrtok led the way along the most nearly level ground. Low creepers became more plentiful, interspersed with scrubby thickets of tangled, spike-armored bushes. Occasionally, small flying things flickered among the foliage. Once, a shrub puffed out an enormous cloud of tiny spores. "Be a job to find anything edible here," grunted Ammet, and Kolin agreed. Finally, after a longer hike than he had anticipated, they approached the edge of the deceptively distant forest. Yrtok paused to examine some purple berries glistening dangerously on a low shrub. Kolin regarded the trees with misgiving. "Looks as tough to get through as a tropical jungle," he remarked. "I think the stuff puts out shoots that grow back into the ground to root as they spread," said the woman. "Maybe we can find a way through." In two or three minutes, they reached the abrupt border of the odd-looking trees. Except for one thick trunked giant, all of them were about the same height. They craned their necks to estimate the altitude of the monster, but the top was hidden by the wide spread of branches. The depths behind it looked dark and impenetrable. "We'd better explore along the edge," decided Yrtok. "Ammet, now is the time to go back and tell the Chief which way we're— Ammet! " Kolin looked over his shoulder. Fifty meters away, Ammet sat beside the bush with the purple berries, utterly relaxed. "He must have tasted some!" exclaimed Kolin. "I'll see how he is." He ran back to the cook and shook him by the shoulder. Ammet's head lolled loosely to one side. His rather heavy features were vacant, lending him a doped appearance. Kolin straightened up and beckoned to Yrtok. For some reason, he had trouble attracting her attention. Then he noticed that she was kneeling. "Hope she didn't eat some stupid thing too!" he grumbled, trotting back. As he reached her, whatever Yrtok was examining came to life and scooted into the underbrush with a flash of greenish fur. All Kolin saw was that it had several legs too many. He pulled Yrtok to her feet. She pawed at him weakly, eyes as vacant as Ammet's. When he let go in sudden horror, she folded gently to the ground. She lay comfortably on her side, twitching one hand as if to brush something away. When she began to smile dreamily, Kolin backed away. The corners of his mouth felt oddly stiff; they had involuntarily drawn back to expose his clenched teeth. He glanced warily about, but nothing appeared to threaten him. "It's time to end this scout," he told himself. "It's dangerous. One good look and I'm jetting off! What I need is an easy tree to climb." He considered the massive giant. Soaring thirty or forty meters into the thin fog and dwarfing other growth, it seemed the most promising choice. At first, Kolin saw no way, but then the network of vines clinging to the rugged trunk suggested a route. He tried his weight gingerly, then began to climb. "I should have brought Yrtok's radio," he muttered. "Oh, well, I can take it when I come down, if she hasn't snapped out of her spell by then. Funny … I wonder if that green thing bit her." Footholds were plentiful among the interlaced lianas. Kolin progressed rapidly. When he reached the first thick limbs, twice head height, he felt safer. Later, at what he hoped was the halfway mark, he hooked one knee over a branch and paused to wipe sweat from his eyes. Peering down, he discovered the ground to be obscured by foliage. "I should have checked from down there to see how open the top is," he mused. "I wonder how the view will be from up there?" "Depends on what you're looking for, Sonny!" something remarked in a soughing wheeze. Kolin, slipping, grabbed desperately for the branch. His fingers clutched a handful of twigs and leaves, which just barely supported him until he regained a grip with the other hand. The branch quivered resentfully under him. "Careful, there!" whooshed the eerie voice. "It took me all summer to grow those!" Kolin could feel the skin crawling along his backbone. "Who are you?" he gasped. The answering sigh of laughter gave him a distinct chill despite its suggestion of amiability. "Name's Johnny Ashlew. Kinda thought you'd start with what I am. Didn't figure you'd ever seen a man grown into a tree before." Kolin looked about, seeing little but leaves and fog. "I have to climb down," he told himself in a reasonable tone. "It's bad enough that the other two passed out without me going space happy too." "What's your hurry?" demanded the voice. "I can talk to you just as easy all the way down, you know. Airholes in my bark—I'm not like an Earth tree." Kolin examined the bark of the crotch in which he sat. It did seem to have assorted holes and hollows in its rough surface. "I never saw an Earth tree," he admitted. "We came from Haurtoz." "Where's that? Oh, never mind—some little planet. I don't bother with them all, since I came here and found out I could be anything I wanted." "What do you mean, anything you wanted?" asked Kolin, testing the firmness of a vertical vine. "Just what I said," continued the voice, sounding closer in his ear as his cheek brushed the ridged bark of the tree trunk. "And, if I do have to remind you, it would be nicer if you said 'Mr. Ashlew,' considering my age." "Your age? How old—?" "Can't really count it in Earth years any more. Lost track. I always figured bein' a tree was a nice, peaceful life; and when I remembered how long some of them live, that settled it. Sonny, this world ain't all it looks like." "It isn't, Mr. Ashlew?" asked Kolin, twisting about in an effort to see what the higher branches might hide. "Nope. Most everything here is run by the Life—that is, by the thing that first grew big enough to do some thinking, and set its roots down all over until it had control. That's the outskirts of it down below." "The other trees? That jungle?" "It's more'n a jungle, Sonny. When I landed here, along with the others from the Arcturan Spark , the planet looked pretty empty to me, just like it must have to—Watch it, there, Boy! If I didn't twist that branch over in time, you'd be bouncing off my roots right now!" "Th-thanks!" grunted Kolin, hanging on grimly. "Doggone vine!" commented the windy whisper. " He ain't one of my crowd. Landed years later in a ship from some star towards the center of the galaxy. You should have seen his looks before the Life got in touch with his mind and set up a mental field to help him change form. He looks twice as good as a vine!" "He's very handy," agreed Kolin politely. He groped for a foothold. "Well … matter of fact, I can't get through to him much, even with the Life's mental field helping. Guess he started living with a different way of thinking. It burns me. I thought of being a tree, and then he came along to take advantage of it!" Kolin braced himself securely to stretch tiring muscles. "Maybe I'd better stay a while," he muttered. "I don't know where I am." "You're about fifty feet up," the sighing voice informed him. "You ought to let me tell you how the Life helps you change form. You don't have to be a tree." "No?" " Uh -uh! Some of the boys that landed with me wanted to get around and see things. Lots changed to animals or birds. One even stayed a man—on the outside anyway. Most of them have to change as the bodies wear out, which I don't, and some made bad mistakes tryin' to be things they saw on other planets." "I wouldn't want to do that, Mr. Ashlew." "There's just one thing. The Life don't like taking chances on word about this place gettin' around. It sorta believes in peace and quiet. You might not get back to your ship in any form that could tell tales." "Listen!" Kolin blurted out. "I wasn't so much enjoying being what I was that getting back matters to me!" "Don't like your home planet, whatever the name was?" "Haurtoz. It's a rotten place. A Planetary State! You have to think and even look the way that's standard thirty hours a day, asleep or awake. You get scared to sleep for fear you might dream treason and they'd find out somehow." "Whooeee! Heard about them places. Must be tough just to live." Suddenly, Kolin found himself telling the tree about life on Haurtoz, and of the officially announced threats to the Planetary State's planned expansion. He dwelt upon the desperation of having no place to hide in case of trouble with the authorities. A multiple system of such worlds was agonizing to imagine. Somehow, the oddity of talking to a tree wore off. Kolin heard opinions spouting out which he had prudently kept bottled up for years. The more he talked and stormed and complained, the more relaxed he felt. "If there was ever a fellow ready for this planet," decided the tree named Ashlew, "you're it, Sonny! Hang on there while I signal the Life by root!" Kolin sensed a lack of direct attention. The rustle about him was natural, caused by an ordinary breeze. He noticed his hands shaking. "Don't know what got into me, talking that way to a tree," he muttered. "If Yrtok snapped out of it and heard, I'm as good as re-personalized right now." As he brooded upon the sorry choice of arousing a search by hiding where he was or going back to bluff things out, the tree spoke. "Maybe you're all set, Sonny. The Life has been thinkin' of learning about other worlds. If you can think of a safe form to jet off in, you might make yourself a deal. How'd you like to stay here?" "I don't know," said Kolin. "The penalty for desertion—" "Whoosh! Who'd find you? You could be a bird, a tree, even a cloud." Silenced but doubting, Kolin permitted himself to try the dream on for size. He considered what form might most easily escape the notice of search parties and still be tough enough to live a long time without renewal. Another factor slipped into his musings: mere hope of escape was unsatisfying after the outburst that had defined his fuming hatred for Haurtoz. I'd better watch myself! he thought. Don't drop diamonds to grab at stars! "What I wish I could do is not just get away but get even for the way they make us live … the whole damn set-up. They could just as easy make peace with the Earth colonies. You know why they don't?" "Why?" wheezed Ashlew. "They're scared that without talk of war, and scouting for Earth fleets that never come, people would have time to think about the way they have to live and who's running things in the Planetary State. Then the gravy train would get blown up—and I mean blown up!" The tree was silent for a moment. Kolin felt the branches stir meditatively. Then Ashlew offered a suggestion. "I could tell the Life your side of it," he hissed. "Once in with us, you can always make thinking connections, no matter how far away. Maybe you could make a deal to kill two birds with one stone, as they used to say on Earth…." Chief Steward Slichow paced up and down beside the ration crate turned up to serve him as a field desk. He scowled in turn, impartially, at his watch and at the weary stewards of his headquarters detail. The latter stumbled about, stacking and distributing small packets of emergency rations. The line of crewmen released temporarily from repair work was transient as to individuals but immutable as to length. Slichow muttered something profane about disregard of orders as he glared at the rocky ridges surrounding the landing place. He was so intent upon planning greetings with which to favor the tardy scouting parties that he failed to notice the loose cloud drifting over the ridge. It was tenuous, almost a haze. Close examination would have revealed it to be made up of myriads of tiny spores. They resembled those cast forth by one of the bushes Kolin's party had passed. Along the edges, the haze faded raggedly into thin air, but the units evidently formed a cohesive body. They drifted together, approaching the men as if taking intelligent advantage of the breeze. One of Chief Slichow's staggering flunkies, stealing a few seconds of relaxation on the pretext of dumping an armful of light plastic packing, wandered into the haze. He froze. After a few heartbeats, he dropped the trash and stared at ship and men as if he had never seen either. A hail from his master moved him. "Coming, Chief!" he called but, returning at a moderate pace, he murmured, "My name is Frazer. I'm a second assistant steward. I'll think as Unit One." Throughout the cloud of spores, the mind formerly known as Peter Kolin congratulated itself upon its choice of form. Nearer to the original shape of the Life than Ashlew got , he thought. He paused to consider the state of the tree named Ashlew, half immortal but rooted to one spot, unable to float on a breeze or through space itself on the pressure of light. Especially, it was unable to insinuate any part of itself into the control center of another form of life, as a second spore was taking charge of the body of Chief Slichow at that very instant. There are not enough men , thought Kolin. Some of me must drift through the airlock. In space, I can spread through the air system to the command group. Repairs to the Peace State and the return to Haurtoz passed like weeks to some of the crew but like brief moments in infinity to other units. At last, the ship parted the air above Headquarters City and landed. The unit known as Captain Theodor Kessel hesitated before descending the ramp. He surveyed the field, the city and the waiting team of inspecting officers. "Could hardly be better, could it?" he chuckled to the companion unit called Security Officer Tarth. "Hardly, sir. All ready for the liberation of Haurtoz." "Reformation of the Planetary State," mused the captain, smiling dreamily as he grasped the handrail. "And then—formation of the Planetary Mind!" END Transcriber's Note: This e-text was produced from Worlds of If January 1962 . Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed. | B. He was intrigued but wanted to try something slightly different. |
Which is true about the role of trust in computing?
A. Increased trust in computers and people is what allowed AT&T to rise in its day
B. Adding more types of computing system reflects an increase in the trust that higher-ups have in their employees to innovate
C. Adding more signposts for trust and approval in computing systems reflects a decrease in trust in their users
D. Increased trust in computers allows for more components of systems to be automated than before
| COMPLEXITY AND HUMANITY We have all seen the images. Volunteers pitching in. People working day and night; coming up with the most ingenious, improvised solutions to everything from food and shelter to communications and security. Working together; patching up the fabric that is rent. Disaster, natural or otherwise, is a breakdown of systems. For a time, chaos reigns. For a time, what will happen in the next five minutes, five hours, and five days is unknown. All we have to rely on are our wits, fortitude, and common humanity Contemporary life is not chaotic, in the colloquial sense we apply to disaster zones. It is, however, complex and rapidly changing; much more so than life was in the past; even the very near past. Life, of course, was never simple. But the fact that day-to-day behaviors in Shenzhen and Bangalore have direct and immediate effects on people from Wichita to Strasbourg, from Rio de Janeiro to Sydney, or that unscrupulous lenders and careless borrowers in the United States can upend economic expectations everywhere else in the world, no matter how carefully others have planned, means that there are many more moving parts that affect each other. And from this scale of practical effects, complexity emerges. New things too were ever under the sun; but the systematic application of knowledge to the creation of new knowledge, innovation to innovation, and information to making more information has become pervasive; and with it the knowledge that next year will be very different than this. The Web, after all, is less than a generation old. These two features−the global scale of interdependence of human action, and the systematic acceleration of innovation, make contemporary life a bit like a slow motion disaster, in one important respect. Its very unpredictability makes it unwise to build systems that take too much away from what human beings do best: look, think, innovate, adapt, discuss, learn, and repeat. That is why we have seen many more systems take on a loose, human centric model in the last decade and a half: from the radical divergence of Toyota’s production system from the highly structured model put in place by Henry Ford, to the Internet’s radical departure from the AT&T system that preceded it, and on to the way Wikipedia constructs human knowledge on the fly, incrementally, in ways that would have been seen, until recently, as too chaotic ever to work (and are still seen so be many). But it is time we acknowledge that systems work best by making work human. Modern Times Modern times were hard enough. Trains and planes, telegraph and telephone, all brought many people into the same causal space. The solution to this increased complexity in the late 19th, early 20th century was to increase the role of structure and improve its design. During the first two-thirds of the twentieth century, this type of rationalization took the form of ever-more complex managed systems, with crisp specification of roles, lines of authority, communication and control. In business, this rationalization was typified by Fredrick Taylor’s Scientific Management, later embodied in Henry Ford’s assembly line. The ambition of these approaches was to specify everything that needed doing in minute detail, to enforce it through monitoring and rewards, and later to build it into the very technology of work−the assembly line. The idea was to eliminate human error and variability in the face of change by removing thinking to the system, and thus neutralizing the variability of the human beings who worked it. Few images captured that time, and what it did to humanity, more vividly than Charlie Chaplin’s assembly line worker in Modern Times. At the same time, government experienced the rise of bureaucratization and the administrative state. Nowhere was this done more brutally than in the totalitarian states of mid-century. But the impulse to build fully-specified systems, designed by experts, monitored and controlled so as to limit human greed and error and to manage uncertainty, was basic and widespread. It underlay the development of the enormously successful state bureaucracies that responded to the Great Depression with the New Deal. It took shape in the Marshall Plan to pull Europe out of the material abyss into which it had been plunged by World War II, and shepherded Japan’s industrial regeneration from it. In technical systems too, we saw in mid-century marvels like the AT&T telephone system and the IBM mainframe. For a moment in history, these large scale managed systems were achieving efficiencies that seemed to overwhelm competing models: from the Tennessee Valley Authority to Sputnik, from Watson’s IBM to General Motors. Yet, to list these paragons from today’s perspective is already to presage the demise of the belief in their inevitable victory. The increasing recognition of the limits of command-and-control systems led to a new approach; but it turned out to be a retrenchment, not an abandonment, of the goal of perfect rationalization of systems design, which assumed much of the human away. What replaced planning and control in these systems was the myth of perfect markets. This was achieved through a hyper-simplification of human nature, wedded to mathematical modeling of what hyper-simplified selfish rational actors, looking only to their own interests, would do under diverse conditions. This approach was widespread and influential; it still is. And yet it led to such unforgettable gems as trying to understand why people do, or do not, use condoms by writing sentences like: “The expected utility (EU) of unsafe sex for m and for f is equal to the benefits (B) of unsafe sex minus its expected costs, and is given by EUm = B - C(1-Pm)(Pf) and EUf = B - C(1-Pf)(Pm),” and believing that you will learn anything useful about lust and desire, recklessness and helplessness, or how to slow down the transmission of AIDS. Only by concocting such a thin model of humanity−no more than the economists’ utility curve−and neglecting any complexities of social interactions that could not be conveyed through prices, could the appearance of rationalization be maintained. Like bureaucratic rationalization, perfect-market rationalization also had successes. But, like its predecessor, its limits as an approach to human systems design are becoming cleare Work, Trust and Play Pricing perfectly requires perfect information. And perfect information, while always an illusion, has become an ever receding dream in a world of constant, rapid change and complex global interactions. What we are seeing instead is the rise of human systems that increasingly shy away from either control or perfect pricing. Not that there isn’t control. Not that there aren’t markets. And not that either of these approaches to coordinating human action will disappear. But these managed systems are becoming increasingly interlaced with looser structures, which invite and enable more engaged human action by drawing on intrinsic motivations and social relations. Dress codes and a culture of play in the workplace in Silicon Valley, like the one day per week that Google employees can use to play at whatever ideas they like, do not exist to make the most innovative region in the United States a Ludic paradise, gratifying employees at the expense of productivity, but rather to engage the human and social in the pursuit of what is, in the long term, the only core business competency−innovation. Wikipedia has eclipsed all the commercial encyclopedias except Britannica not by issuing a large IPO and hiring the smartest guys in the room, but by building an open and inviting system that lets people learn together and pursue their passion for knowledge, and each other’s company. The set of human systems necessary for action in this complex, unpredictable set of conditions, combining rationalization with human agency, learning and adaptation, is as different from managed systems and perfect markets as the new Toyota is from the old General Motors, or as the Internet now is from AT&T then. The hallmarks of these newer systems are: (a) location of authority and practical capacity to act at the edges of the system, where potentialities for sensing the environment, identifying opportunities and challenges to action and acting upon them, are located; (b) an emphasis on the human: on trust, cooperation, judgment and insight; (c) communication over the lifetime of the interaction; and (d) loosely-coupled systems: systems in which the regularities and dependencies among objects and processes are less strictly associated with each other; where actions and interactions can occur through multiple systems simultaneously, have room to fail, maneuver, and be reoriented to fit changing conditions and new learning, or shift from one system to another to achieve a solution. Consider first of all the triumph of Toyota over the programs of Taylor and Ford. Taylorism was typified by the ambition to measure and specify all human and material elements of the production system. The ambition of scientific management was to offer a single, integrated system where all human variance (the source of slothful shirking and inept error) could be isolated and controlled. Fordism took that ambition and embedded the managerial knowledge in the technological platform of the assembly line, guided by a multitude of rigid task specifications and routines. Toyota Production System, by comparison, has a substantially smaller number of roles that are also more loosely defined, with a reliance on small teams where each team member can perform all tasks, and who are encouraged to experiment, improve, fail, adapt, but above all communicate. The system is built on trust and a cooperative dynamic. The enterprise functions through a managerial control system, but also through social cooperation mechanisms built around teamwork and trust. However, even Toyota might be bested in this respect by the even more loosely coupled networks of innovation and supply represented by Taiwanese original-design manufacturers. But let us also consider the system in question that has made this work possible, the Internet, and compare it to the design principles of the AT&T network in its heyday. Unlike the Internet, AT&T’s network was fully managed. Mid-century, the company even retained ownership of the phones at the endpoints, arguing that it needed to prohibit customers from connecting unlicensed phones to the system (ostensibly to ensure proper functioning of the networking and monitoring of customer behavior, although it didn’t hurt either that this policy effectively excluded competitors). This generated profit, but any substantial technical innovations required the approval of management and a re-engineering of the entire network. The Internet, on the other hand, was designed to be as general as possible. The network hardware merely delivers packets of data using standardized addressing information. The hard processing work−manipulating a humanly-meaningful communication (a letter or a song, a video or a software package) and breaking it up into a stream of packets−was to be done by its edge devices, in this case computers owned by users. This system allowed the breathtaking rate of innovation that we have seen, while also creating certain vulnerabilities in online security. These vulnerabilities have led some to argue that a new system to manage the Internet is needed. We see first of all that doubts about trust and security on the Internet arise precisely because the network was originally designed for people who could more-or-less trust each other, and offloaded security from the network to the edges. As the network grew and users diversified, trust (the practical belief that other human agents in the system were competent and benign, or at least sincere) declined. This decline was met with arguments in favor of building security into the technical system, both at its core, in the network elements themselves, and at its periphery, through “trusted computing.” A “trusted computer” will, for example, not run a program or document that its owner wants to run, unless it has received authorization from some other locus: be it the copyright owner, the virus protection company, or the employer. This is thought to be the most completely effective means of preventing copyright infringement or system failure, and preserving corporate security (these are the main reasons offered for implementing such systems). Trusted computing in this form is the ultimate reversal of the human-centric, loosely-coupled design approach of the Internet. Instead of locating authority and capacity to act at the endpoints, where human beings are located and can make decisions about what is worthwhile, it implements the belief that machines−technical systems−are trustworthy, while their human users are malevolent, incompetent, or both. Reintroducing the Human Taylorism, the Bell system and trusted computing are all efforts to remove human agency from action and replace it with well-designed, tightly-bound systems. That is, the specifications and regularities of the system are such that they control or direct action and learning over time. Human agency, learning, communication and adaptation are minimized in managed systems, if not eliminated, and the knowledge in the system comes from the outside, from the designer, in the initial design over time, and through observation of the system’s performance by someone standing outside its constraints−a manager or systems designer. By contrast, loosely-coupled systems affirmatively eschew this level of control, and build in room for human agency, experimentation, failure, communication, learning and adaptation. Loose-coupling is central to the new systems. It is a feature of system design that leaves room for human agency over time, only imperfectly constraining and enabling any given action by the system itself. By creating such domains of human agency, system designers are accepting the limitations of design and foresight, and building in the possibilities of learning over time through action in the system, by agents acting within To deal with the new complexity of contemporary life we need to re-introduce the human into the design of systems. We must put the soul back into the system. If years of work on artificial intelligence have taught us anything, it is that what makes for human insight is extremely difficult to replicate or systematize. At the center of these new systems, then, sits a human being who has a capacity to make judgments, experiment, learn and adapt. But enabling human agency also provides scope of action for human frailty. Although this idea is most alien to the mainstream of system design in the twentieth century, we must now turn our attention to building systems that support human sociality−our ability to think of others and their needs, and to choose for ourselves goals consistent with a broader social concern than merely our own self-interest. The challenge of the near future is to build systems that will allow us to be largely free to inquire, experiment, learn and communicate, that will encourage us to cooperate, and that will avoid the worst of what human beings are capable of, and elicit what is best. Free software, Wikipedia, Creative Commons and the thousands of emerging human practices of productive social cooperation in the networked information economy give us real existence proofs that human-centric systems can not merely exist, but thrive, as can the human beings and social relations that make them. | C. Adding more signposts for trust and approval in computing systems reflects a decrease in trust in their users |
What would have happened if Ledman had stayed on Earth?
A. He would've managed to maintain leadership of his company
B. He might not have needed a wheelchair long-term
C. He would have joined the Project Sea-Dredge mission
D. He would've become more depressed and never found revenge
| THE HUNTED HEROES By ROBERT SILVERBERG The planet itself was tough enough—barren, desolate, forbidding; enough to stop the most adventurous and dedicated. But they had to run head-on against a mad genius who had a motto: Death to all Terrans! "Let's keep moving," I told Val. "The surest way to die out here on Mars is to give up." I reached over and turned up the pressure on her oxymask to make things a little easier for her. Through the glassite of the mask, I could see her face contorted in an agony of fatigue. And she probably thought the failure of the sandcat was all my fault, too. Val's usually about the best wife a guy could ask for, but when she wants to be she can be a real flying bother. It was beyond her to see that some grease monkey back at the Dome was at fault—whoever it was who had failed to fasten down the engine hood. Nothing but what had stopped us could stop a sandcat: sand in the delicate mechanism of the atomic engine. But no; she blamed it all on me somehow: So we were out walking on the spongy sand of the Martian desert. We'd been walking a good eight hours. "Can't we turn back now, Ron?" Val pleaded. "Maybe there isn't any uranium in this sector at all. I think we're crazy to keep on searching out here!" I started to tell her that the UranCo chief had assured me we'd hit something out this way, but changed my mind. When Val's tired and overwrought there's no sense in arguing with her. I stared ahead at the bleak, desolate wastes of the Martian landscape. Behind us somewhere was the comfort of the Dome, ahead nothing but the mazes and gullies of this dead world. He was a cripple in a wheelchair—helpless as a rattlesnake. "Try to keep going, Val." My gloved hand reached out and clumsily enfolded hers. "Come on, kid. Remember—we're doing this for Earth. We're heroes." She glared at me. "Heroes, hell!" she muttered. "That's the way it looked back home, but, out there it doesn't seem so glorious. And UranCo's pay is stinking." "We didn't come out here for the pay, Val." "I know, I know, but just the same—" It must have been hell for her. We had wandered fruitlessly over the red sands all day, both of us listening for the clicks of the counter. And the geigers had been obstinately hushed all day, except for their constant undercurrent of meaningless noises. Even though the Martian gravity was only a fraction of Earth's, I was starting to tire, and I knew it must have been really rough on Val with her lovely but unrugged legs. "Heroes," she said bitterly. "We're not heroes—we're suckers! Why did I ever let you volunteer for the Geig Corps and drag me along?" Which wasn't anywhere close to the truth. Now I knew she was at the breaking point, because Val didn't lie unless she was so exhausted she didn't know what she was doing. She had been just as much inflamed by the idea of coming to Mars to help in the search for uranium as I was. We knew the pay was poor, but we had felt it a sort of obligation, something we could do as individuals to keep the industries of radioactives-starved Earth going. And we'd always had a roving foot, both of us. No, we had decided together to come to Mars—the way we decided together on everything. Now she was turning against me. I tried to jolly her. "Buck up, kid," I said. I didn't dare turn up her oxy pressure any higher, but it was obvious she couldn't keep going. She was almost sleep-walking now. We pressed on over the barren terrain. The geiger kept up a fairly steady click-pattern, but never broke into that sudden explosive tumult that meant we had found pay-dirt. I started to feel tired myself, terribly tired. I longed to lie down on the soft, spongy Martian sand and bury myself. I looked at Val. She was dragging along with her eyes half-shut. I felt almost guilty for having dragged her out to Mars, until I recalled that I hadn't. In fact, she had come up with the idea before I did. I wished there was some way of turning the weary, bedraggled girl at my side back into the Val who had so enthusiastically suggested we join the Geigs. Twelve steps later, I decided this was about as far as we could go. I stopped, slipped out of the geiger harness, and lowered myself ponderously to the ground. "What'samatter, Ron?" Val asked sleepily. "Something wrong?" "No, baby," I said, putting out a hand and taking hers. "I think we ought to rest a little before we go any further. It's been a long, hard day." It didn't take much to persuade her. She slid down beside me, curled up, and in a moment she was fast asleep, sprawled out on the sands. Poor kid , I thought. Maybe we shouldn't have come to Mars after all. But, I reminded myself, someone had to do the job. A second thought appeared, but I squelched it: Why the hell me? I looked down at Valerie's sleeping form, and thought of our warm, comfortable little home on Earth. It wasn't much, but people in love don't need very fancy surroundings. I watched her, sleeping peacefully, a wayward lock of her soft blonde hair trailing down over one eyebrow, and it seemed hard to believe that we'd exchanged Earth and all it held for us for the raw, untamed struggle that was Mars. But I knew I'd do it again, if I had the chance. It's because we wanted to keep what we had. Heroes? Hell, no. We just liked our comforts, and wanted to keep them. Which took a little work. Time to get moving. But then Val stirred and rolled over in her sleep, and I didn't have the heart to wake her. I sat there, holding her, staring out over the desert, watching the wind whip the sand up into weird shapes. The Geig Corps preferred married couples, working in teams. That's what had finally decided it for us—we were a good team. We had no ties on Earth that couldn't be broken without much difficulty. So we volunteered. And here we are. Heroes. The wind blasted a mass of sand into my face, and I felt it tinkle against the oxymask. I glanced at the suit-chronometer. Getting late. I decided once again to wake Val. But she was tired. And I was tired too, tired from our wearying journey across the empty desert. I started to shake Val. But I never finished. It would be so nice just to lean back and nuzzle up to her, down in the sand. So nice. I yawned, and stretched back. I awoke with a sudden startled shiver, and realized angrily I had let myself doze off. "Come on, Val," I said savagely, and started to rise to my feet. I couldn't. I looked down. I was neatly bound in thin, tough, plastic tangle-cord, swathed from chin to boot-bottoms, my arms imprisoned, my feet caught. And tangle-cord is about as easy to get out of as a spider's web is for a trapped fly. It wasn't Martians that had done it. There weren't any Martians, hadn't been for a million years. It was some Earthman who had bound us. I rolled my eyes toward Val, and saw that she was similarly trussed in the sticky stuff. The tangle-cord was still fresh, giving off a faint, repugnant odor like that of drying fish. It had been spun on us only a short time ago, I realized. "Ron—" "Don't try to move, baby. This stuff can break your neck if you twist it wrong." She continued for a moment to struggle futilely, and I had to snap, "Lie still, Val!" "A very wise statement," said a brittle, harsh voice from above me. I looked up and saw a helmeted figure above us. He wasn't wearing the customary skin-tight pliable oxysuits we had. He wore an outmoded, bulky spacesuit and a fishbowl helmet, all but the face area opaque. The oxygen cannisters weren't attached to his back as expected, though. They were strapped to the back of the wheelchair in which he sat. Through the fishbowl I could see hard little eyes, a yellowed, parchment-like face, a grim-set jaw. I didn't recognize him, and this struck me odd. I thought I knew everyone on sparsely-settled Mars. Somehow I'd missed him. What shocked me most was that he had no legs. The spacesuit ended neatly at the thighs. He was holding in his left hand the tanglegun with which he had entrapped us, and a very efficient-looking blaster was in his right. "I didn't want to disturb your sleep," he said coldly. "So I've been waiting here for you to wake up." I could just see it. He might have been sitting there for hours, complacently waiting to see how we'd wake up. That was when I realized he must be totally insane. I could feel my stomach-muscles tighten, my throat constrict painfully. Then anger ripped through me, washing away the terror. "What's going on?" I demanded, staring at the half of a man who confronted us from the wheelchair. "Who are you?" "You'll find out soon enough," he said. "Suppose now you come with me." He reached for the tanglegun, flipped the little switch on its side to MELT, and shot a stream of watery fluid over our legs, keeping the blaster trained on us all the while. Our legs were free. "You may get up now," he said. "Slowly, without trying to make trouble." Val and I helped each other to our feet as best we could, considering our arms were still tightly bound against the sides of our oxysuits. "Walk," the stranger said, waving the tanglegun to indicate the direction. "I'll be right behind you." He holstered the tanglegun. I glimpsed the bulk of an outboard atomic rigging behind him, strapped to the back of the wheelchair. He fingered a knob on the arm of the chair and the two exhaust ducts behind the wheel-housings flamed for a moment, and the chair began to roll. Obediently, we started walking. You don't argue with a blaster, even if the man pointing it is in a wheelchair. "What's going on, Ron?" Val asked in a low voice as we walked. Behind us the wheelchair hissed steadily. "I don't quite know, Val. I've never seen this guy before, and I thought I knew everyone at the Dome." "Quiet up there!" our captor called, and we stopped talking. We trudged along together, with him following behind; I could hear the crunch-crunch of the wheelchair as its wheels chewed into the sand. I wondered where we were going, and why. I wondered why we had ever left Earth. The answer to that came to me quick enough: we had to. Earth needed radioactives, and the only way to get them was to get out and look. The great atomic wars of the late 20th Century had used up much of the supply, but the amount used to blow up half the great cities of the world hardly compared with the amount we needed to put them back together again. In three centuries the shattered world had been completely rebuilt. The wreckage of New York and Shanghai and London and all the other ruined cities had been hidden by a shining new world of gleaming towers and flying roadways. We had profited by our grandparents' mistakes. They had used their atomics to make bombs. We used ours for fuel. It was an atomic world. Everything: power drills, printing presses, typewriters, can openers, ocean liners, powered by the inexhaustible energy of the dividing atom. But though the energy is inexhaustible, the supply of nuclei isn't. After three centuries of heavy consumption, the supply failed. The mighty machine that was Earth's industry had started to slow down. And that started the chain of events that led Val and me to end up as a madman's prisoners, on Mars. With every source of uranium mined dry on Earth, we had tried other possibilities. All sorts of schemes came forth. Project Sea-Dredge was trying to get uranium from the oceans. In forty or fifty years, they'd get some results, we hoped. But there wasn't forty or fifty years' worth of raw stuff to tide us over until then. In a decade or so, our power would be just about gone. I could picture the sort of dog-eat-dog world we'd revert back to. Millions of starving, freezing humans tooth-and-clawing in it in the useless shell of a great atomic civilization. So, Mars. There's not much uranium on Mars, and it's not easy to find or any cinch to mine. But what little is there, helps. It's a stopgap effort, just to keep things moving until Project Sea-Dredge starts functioning. Enter the Geig Corps: volunteers out on the face of Mars, combing for its uranium deposits. And here we are, I thought. After we walked on a while, a Dome became visible up ahead. It slid up over the crest of a hill, set back between two hummocks on the desert. Just out of the way enough to escape observation. For a puzzled moment I thought it was our Dome, the settlement where all of UranCo's Geig Corps were located, but another look told me that this was actually quite near us and fairly small. A one-man Dome, of all things! "Welcome to my home," he said. "The name is Gregory Ledman." He herded us off to one side of the airlock, uttered a few words keyed to his voice, and motioned us inside when the door slid up. When we were inside he reached up, clumsily holding the blaster, and unscrewed the ancient spacesuit fishbowl. His face was a bitter, dried-up mask. He was a man who hated. The place was spartanly furnished. No chairs, no tape-player, no decoration of any sort. Hard bulkhead walls, rivet-studded, glared back at us. He had an automatic chef, a bed, and a writing-desk, and no other furniture. Suddenly he drew the tanglegun and sprayed our legs again. We toppled heavily to the floor. I looked up angrily. "I imagine you want to know the whole story," he said. "The others did, too." Valerie looked at me anxiously. Her pretty face was a dead white behind her oxymask. "What others?" "I never bothered to find out their names," Ledman said casually. "They were other Geigs I caught unawares, like you, out on the desert. That's the only sport I have left—Geig-hunting. Look out there." He gestured through the translucent skin of the Dome, and I felt sick. There was a little heap of bones lying there, looking oddly bright against the redness of the sands. They were the dried, parched skeletons of Earthmen. Bits of cloth and plastic, once oxymasks and suits, still clung to them. Suddenly I remembered. There had been a pattern there all the time. We didn't much talk about it; we chalked it off as occupational hazards. There had been a pattern of disappearances on the desert. I could think of six, eight names now. None of them had been particularly close friends. You don't get time to make close friends out here. But we'd vowed it wouldn't happen to us. It had. "You've been hunting Geigs?" I asked. " Why? What've they ever done to you?" He smiled, as calmly as if I'd just praised his house-keeping. "Because I hate you," he said blandly. "I intend to wipe every last one of you out, one by one." I stared at him. I'd never seen a man like this before; I thought all his kind had died at the time of the atomic wars. I heard Val sob, "He's a madman!" "No," Ledman said evenly. "I'm quite sane, believe me. But I'm determined to drive the Geigs—and UranCo—off Mars. Eventually I'll scare you all away." "Just pick us off in the desert?" "Exactly," replied Ledman. "And I have no fears of an armed attack. This place is well fortified. I've devoted years to building it. And I'm back against those hills. They couldn't pry me out." He let his pale hand run up into his gnarled hair. "I've devoted years to this. Ever since—ever since I landed here on Mars." "What are you going to do with us?" Val finally asked, after a long silence. He didn't smile this time. "Kill you," he told her. "Not your husband. I want him as an envoy, to go back and tell the others to clear off." He rocked back and forth in his wheelchair, toying with the gleaming, deadly blaster in his hand. We stared in horror. It was a nightmare—sitting there, placidly rocking back and forth, a nightmare. I found myself fervently wishing I was back out there on the infinitely safer desert. "Do I shock you?" he asked. "I shouldn't—not when you see my motives." "We don't see them," I snapped. "Well, let me show you. You're on Mars hunting uranium, right? To mine and ship the radioactives back to Earth to keep the atomic engines going. Right?" I nodded over at our geiger counters. "We volunteered to come to Mars," Val said irrelevantly. "Ah—two young heroes," Ledman said acidly. "How sad. I could almost feel sorry for you. Almost." "Just what is it you're after?" I said, stalling, stalling. "Atomics cost me my legs," he said. "You remember the Sadlerville Blast?" he asked. "Of course." And I did, too. I'd never forget it. No one would. How could I forget that great accident—killing hundreds, injuring thousands more, sterilizing forty miles of Mississippi land—when the Sadlerville pile went up? "I was there on business at the time," Ledman said. "I represented Ledman Atomics. I was there to sign a new contract for my company. You know who I am, now?" I nodded. "I was fairly well shielded when it happened. I never got the contract, but I got a good dose of radiation instead. Not enough to kill me," he said. "Just enough to necessitate the removal of—" he indicated the empty space at his thighs. "So I got off lightly." He gestured at the wheelchair blanket. I still didn't understand. "But why kill us Geigs? We had nothing to do with it." "You're just in this by accident," he said. "You see, after the explosion and the amputation, my fellow-members on the board of Ledman Atomics decided that a semi-basket case like myself was a poor risk as Head of the Board, and they took my company away. All quite legal, I assure you. They left me almost a pauper!" Then he snapped the punchline at me. "They renamed Ledman Atomics. Who did you say you worked for?" I began, "Uran—" "Don't bother. A more inventive title than Ledman Atomics, but not quite as much heart, wouldn't you say?" He grinned. "I saved for years; then I came to Mars, lost myself, built this Dome, and swore to get even. There's not a great deal of uranium on this planet, but enough to keep me in a style to which, unfortunately, I'm no longer accustomed." He consulted his wrist watch. "Time for my injection." He pulled out the tanglegun and sprayed us again, just to make doubly certain. "That's another little souvenir of Sadlerville. I'm short on red blood corpuscles." He rolled over to a wall table and fumbled in a container among a pile of hypodermics. "There are other injections, too. Adrenalin, insulin. Others. The Blast turned me into a walking pin-cushion. But I'll pay it all back," he said. He plunged the needle into his arm. My eyes widened. It was too nightmarish to be real. I wasn't seriously worried about his threat to wipe out the entire Geig Corps, since it was unlikely that one man in a wheelchair could pick us all off. No, it wasn't the threat that disturbed me, so much as the whole concept, so strange to me, that the human mind could be as warped and twisted as Ledman's. I saw the horror on Val's face, and I knew she felt the same way I did. "Do you really think you can succeed?" I taunted him. "Really think you can kill every Earthman on Mars? Of all the insane, cockeyed—" Val's quick, worried head-shake cut me off. But Ledman had felt my words, all right. "Yes! I'll get even with every one of you for taking away my legs! If we hadn't meddled with the atom in the first place, I'd be as tall and powerful as you, today—instead of a useless cripple in a wheelchair." "You're sick, Gregory Ledman," Val said quietly. "You've conceived an impossible scheme of revenge and now you're taking it out on innocent people who've done nothing, nothing at all to you. That's not sane!" His eyes blazed. "Who are you to talk of sanity?" Uneasily I caught Val's glance from a corner of my eye. Sweat was rolling down her smooth forehead faster than the auto-wiper could swab it away. "Why don't you do something? What are you waiting for, Ron?" "Easy, baby," I said. I knew what our ace in the hole was. But I had to get Ledman within reach of me first. "Enough," he said. "I'm going to turn you loose outside, right after—" " Get sick! " I hissed to Val, low. She began immediately to cough violently, emitting harsh, choking sobs. "Can't breathe!" She began to yell, writhing in her bonds. That did it. Ledman hadn't much humanity left in him, but there was a little. He lowered the blaster a bit and wheeled one-hand over to see what was wrong with Val. She continued to retch and moan most horribly. It almost convinced me. I saw Val's pale, frightened face turn to me. He approached and peered down at her. He opened his mouth to say something, and at that moment I snapped my leg up hard, tearing the tangle-cord with a snicking rasp, and kicked his wheelchair over. The blaster went off, burning a hole through the Dome roof. The automatic sealers glued-in instantly. Ledman went sprawling helplessly out into the middle of the floor, the wheelchair upended next to him, its wheels slowly revolving in the air. The blaster flew from his hands at the impact of landing and spun out near me. In one quick motion I rolled over and covered it with my body. Ledman clawed his way to me with tremendous effort and tried wildly to pry the blaster out from under me, but without success. I twisted a bit, reached out with my free leg, and booted him across the floor. He fetched up against the wall of the Dome and lay there. Val rolled over to me. "Now if I could get free of this stuff," I said, "I could get him covered before he comes to. But how?" "Teamwork," Val said. She swivelled around on the floor until her head was near my boot. "Push my oxymask off with your foot, if you can." I searched for the clamp and tried to flip it. No luck, with my heavy, clumsy boot. I tried again, and this time it snapped open. I got the tip of my boot in and pried upward. The oxymask came off, slowly, scraping a jagged red scratch up the side of Val's neck as it came. "There," she breathed. "That's that." I looked uneasily at Ledman. He was groaning and beginning to stir. Val rolled on the floor and her face lay near my right arm. I saw what she had in mind. She began to nibble the vile-tasting tangle-cord, running her teeth up and down it until it started to give. She continued unfailingly. Finally one strand snapped. Then another. At last I had enough use of my hand to reach out and grasp the blaster. Then I pulled myself across the floor to Ledman, removed the tanglegun, and melted the remaining tangle-cord off. My muscles were stiff and bunched, and rising made me wince. I turned and freed Val. Then I turned and faced Ledman. "I suppose you'll kill me now," he said. "No. That's the difference between sane people and insane," I told him. "I'm not going to kill you at all. I'm going to see to it that you're sent back to Earth." " No! " he shouted. "No! Anything but back there. I don't want to face them again—not after what they did to me—" "Not so loud," I broke in. "They'll help you on Earth. They'll take all the hatred and sickness out of you, and turn you into a useful member of society again." "I hate Earthmen," he spat out. "I hate all of them." "I know," I said sarcastically. "You're just all full of hate. You hated us so much that you couldn't bear to hang around on Earth for as much as a year after the Sadlerville Blast. You had to take right off for Mars without a moment's delay, didn't you? You hated Earth so much you had to leave." "Why are you telling all this to me?" "Because if you'd stayed long enough, you'd have used some of your pension money to buy yourself a pair of prosthetic legs, and then you wouldn't need this wheelchair." Ledman scowled, and then his face went belligerent again. "They told me I was paralyzed below the waist. That I'd never walk again, even with prosthetic legs, because I had no muscles to fit them to." "You left Earth too quickly," Val said. "It was the only way," he protested. "I had to get off—" "She's right," I told him. "The atom can take away, but it can give as well. Soon after you left they developed atomic-powered prosthetics—amazing things, virtually robot legs. All the survivors of the Sadlerville Blast were given the necessary replacement limbs free of charge. All except you. You were so sick you had to get away from the world you despised and come here." "You're lying," he said. "It's not true!" "Oh, but it is," Val smiled. I saw him wilt visibly, and for a moment I almost felt sorry for him, a pathetic legless figure propped up against the wall of the Dome at blaster-point. But then I remembered he'd killed twelve Geigs—or more—and would have added Val to the number had he had the chance. "You're a very sick man, Ledman," I said. "All this time you could have been happy, useful on Earth, instead of being holed up here nursing your hatred. You might have been useful, on Earth. But you decided to channel everything out as revenge." "I still don't believe it—those legs. I might have walked again. No—no, it's all a lie. They told me I'd never walk," he said, weakly but stubbornly still. I could see his whole structure of hate starting to topple, and I decided to give it the final push. "Haven't you wondered how I managed to break the tangle-cord when I kicked you over?" "Yes—human legs aren't strong enough to break tangle-cord that way." "Of course not," I said. I gave Val the blaster and slipped out of my oxysuit. "Look," I said. I pointed to my smooth, gleaming metal legs. The almost soundless purr of their motors was the only noise in the room. "I was in the Sadlerville Blast, too," I said. "But I didn't go crazy with hate when I lost my legs." Ledman was sobbing. "Okay, Ledman," I said. Val got him into his suit, and brought him the fishbowl helmet. "Get your helmet on and let's go. Between the psychs and the prosthetics men, you'll be a new man inside of a year." "But I'm a murderer!" "That's right. And you'll be sentenced to psych adjustment. When they're finished, Gregory Ledman the killer will be as dead as if they'd electrocuted you, but there'll be a new—and sane—Gregory Ledman." I turned to Val. "Got the geigers, honey?" For the first time since Ledman had caught us, I remembered how tired Val had been out on the desert. I realized now that I had been driving her mercilessly—me, with my chromium legs and atomic-powered muscles. No wonder she was ready to fold! And I'd been too dense to see how unfair I had been. She lifted the geiger harnesses, and I put Ledman back in his wheelchair. Val slipped her oxymask back on and fastened it shut. "Let's get back to the Dome in a hurry," I said. "We'll turn Ledman over to the authorities. Then we can catch the next ship for Earth." "Go back? Go back? If you think I'm backing down now and quitting you can find yourself another wife! After we dump this guy I'm sacking in for twenty hours, and then we're going back out there to finish that search-pattern. Earth needs uranium, honey, and I know you'd never be happy quitting in the middle like that." She smiled. "I can't wait to get out there and start listening for those tell-tale clicks." I gave a joyful whoop and swung her around. When I put her down, she squeezed my hand, hard. "Let's get moving, fellow hero," she said. I pressed the stud for the airlock, smiling. THE END Transcriber's Note: This etext was produced from Amazing Stories September 1956. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed. Minor spelling and typographical errors have been corrected without note. | B. He might not have needed a wheelchair long-term |
What do neurons do?
A. Form vastly interconnected networks
B. Process information and learn
C. Change the strength of the synapses between cells
D. Share chemical signals with neighboring cells
| Sharism: A Mind Revolution With the People of the World Wide Web communicating more fully and freely in Social Media while rallying a Web 2.0 content boom, the inner dynamics of such a creative explosion must be studied more closely. What motivates those who join this movement and what future will they create? A key fact is that a superabundance of community respect and social capital are being accumulated by those who share. The key motivator of Social Media and the core spirit of Web 2.0 is a mind switch called Sharism. Sharism suggests a re-orientation of personal values. We see it in User Generated Content. It is the pledge of Creative Commons. It is in the plans of future-oriented cultural initiatives. Sharism is also a mental practice that anyone can try, a social-psychological attitude to transform a wide and isolated world into a super-smart Social Brain. The Neuron Doctrine Sharism is encoded in the Human Genome. Although eclipsed by the many pragmatisms of daily life, the theory of Sharism finds basis in neuroscience and its study of the working model of the human brain. Although we can’t entirely say how the brain works as a whole, we do have a model of the functional mechanism of the nervous system and its neurons. A neuron is not a simple organic cell, but a very powerful, electrically excitable biological processor. Groups of neurons form vastly interconnected networks, which, by changing the strength of the synapses between cells, can process information, and learn. A neuron, by sharing chemical signals with its neighbors, can be integrated into more meaningful patterns that keep the neuron active and alive. Moreover, such a simple logic can be iterated and amplified, since all neurons work on a similar principle of connecting and sharing. Originally, the brain is quite open. A neural network exists to share activity and information, and I believe this model of the brain should inspire ideas and decisions about human networks. Thus, our brain supports sharing in its very system-nature. This has profound implications for the creative process. Whenever you have an intention to create, you will find it easier to generate more creative ideas if you keep the sharing process firmly in mind. The idea-forming-process is not linear, but more like an avalanche of amplifications along the thinking path. It moves with the momentum of a creative snowball. If your internal cognitive system encourages sharing, you can engineer a feedback loop of happiness, which will help you generate even more ideas in return. It’s a kind of butterfly- effect, as the small creative energy you spend will eventually return to make you, and the world, more creative. However, daily decisions for most adults are quite low in creative productivity, if only because they’ve switched off their sharing paths. People generally like to share what they create, but in a culture that tells them to be protective of their ideas, people start to believe in the danger of sharing. Then Sharism will be degraded in their mind and not encouraged in their society. But if we can encourage someone to share, her sharing paths will stay open. Sharism will be kept in her mind as a memory and an instinct. If in the future she faces a creative choice, her choice will be, “Share.” These mind-switches are too subtle to be felt. But since the brain, and society, is a connected system, the accumulation of these micro-attitudes, from neuron to neuron and person to person, can result in observable behavior. It is easy to tell if a person, a group, a company, a nation is oriented toward Sharism or not. For those who are not, what they defend as “cultural goods” and “intellectual property” are just excuses for the status quo of keeping a community closed. Much of their “culture” will be protected, but the net result is the direct loss of many other precious ideas, and the subsequent loss of all the potential gains of sharing. This lost knowledge is a black hole in our life, which may start to swallow other values as well. Non-sharing culture misleads us with its absolute separation of Private and Public space. It makes creative action a binary choice between public and private, open and closed. This creates a gap in the spectrum of knowledge. Although this gap has the potential to become a valuable creative space, concerns about privacy make this gap hard to fill. We shouldn’t be surprised that, to be safe, most people keep their sharing private and stay “closed.” They may fear the Internet creates a potential for abuse that they can’t fight alone. However, the paradox is: The less you share, the less power you have. New Technologies and the Rise of Sharism Let’s track back to 1999, when there were only a few hundred pioneer bloggers around the world, and no more than ten times that many readers following each blog. Human history is always so: something important was happening, but the rest of the world hadn’t yet realized it. The shift toward easy-to-use online publishing triggered a soft revolution in just five years. People made a quick and easy transition from reading blogs, to leaving comments and taking part in online conversations, and then to the sudden realization that they should become bloggers themselves. More bloggers created more readers, and more readers made more blogs. The revolution was viral. Bloggers generate lively and timely information on the Internet, and connect to each other with RSS, hyperlinks, comments, trackbacks and quotes. The small-scale granularity of the content can fill discrete gaps in experience and thus record a new human history. Once you become a blogger, once you have accumulated so much social capital in such a small site, it’s hard to stop. We can’t explain this fact with a theory of addiction. It’s an impulse to share. It’s the energy of the memes that want to be passed from mouth to mouth and mind to mind. It’s more than just E-mail. It’s Sharism. Bloggers are always keen to keep the social context of their posts in mind, by asking themselves, “Who is going to see this?” Bloggers are agile in adjusting their tone−and privacy settings−to advance ideas and stay out of trouble. It’s not self-censorship, but a sense of smart expression. But once blogs reached the tipping point, they expanded into the blogosphere. This required a more delicate social networking system and content- sharing architecture. But people now understand that they can have better control over a wide spectrum of relationships. Like how Flickr allows people to share their photos widely, but safely. The checkbox-based privacy of Flickr may seem unfamiliar to a new user, but you can use it to toy with the mind-switches of Sharism. By checking a box we can choose to share or not to share. From my observations, I have seen photographers on Flickr become more open to sharing, while retaining flexible choices. The rapid emergence of Social Applications that can communicate and cooperate, by allowing people to output content from one service to another, is letting users pump their memes into a pipeline-like ecosystem. This interconnectedness allows memes to travel along multiple online social networks, and potentially reach a huge audience. As a result, such a Micro-pipeline system is making Social Media a true alternative to broadcast media. These new technologies are reviving Sharism in our closed culture. Local Practice, Global Gain If you happened to lose your Sharism in a bad educational or cultural setting, it’s hard to get it back. But it’s not impossible. A persistence of practice can lead to a full recovery. You can think of Sharism as a spiritual practice. But you must practice everyday. Otherwise, you might lose the power of sharing. Permanently. You might need something to spur you on, to keep you from quitting and returning to a closed mindset. Here’s an idea: put a sticky note on your desk that says, “What do you want to share today?” I’m not kidding. Then, if anything interesting comes your way: Share It! The easiest way to both start and keep sharing is by using different kinds of social software applications. Your first meme you want to share may be small, but you can amplify it with new technologies. Enlist some people from your network and invite them into a new social application. At first it might be hard to feel the gains of Sharism. The true test then is to see if you can keep track of the feedback that you get from sharing. You will realize that almost all sharing activities will generate positive results. The happiness that this will obtain is only the most immediate reward. But there are others. The first type of reward that you will get comes in the form of comments. Then you know you’ve provoked interest, appreciation, excitement. The second reward is access to all the other stuff being shared by friends in your network. Since you know and trust them, you will be that much more interested in what they have to share. Already, the return is a multiple of the small meme you first shared. But the third type of return is more dramatic still. Anything you share can be forwarded, circulated and republished via other people’s networks. This cascade effect can spread your work to the networked masses. Improvements in social software are making the speed of dissemination as fast as a mouse-click. You should get to know the Sharism-You. You’re about to become popular, and fast This brings us to the fourth and final type of return. It has a meaning not only for you, but for the whole of society. If you so choose, you may allow others to create derivative works from what you share. This one choice could easily snowball into more creations along the sharing path, from people at key nodes in the network who are all as passionate about creating and sharing as you are. After many iterative rounds of development, a large creative work may spring from your choice to share. Of course, you will get the credit that you asked for, and deserve. And it’s okay to seek financial rewards. But you will in every case get something just as substantial: Happiness. The more people who create in the spirit of Sharism, the easier it will be to attain well- balanced and equitable Social Media that is woven by people themselves. Media won’t be controlled by any single person but will rely on the even distribution of social networking. These “Shaeros” (Sharing Heroes) will naturally become the opinion leaders in the first wave of Social Media. However, these media rights will belong to everyone. You yourself can be both producer and consumer in such a system. Sharism Safeguards Your Rights Still, many questions will be raised about Sharism as an initiative in new age. The main one is copyright. One concern is that any loss of control over copyrighted content will lead to noticeable deficits in personal wealth, or just loss of control. 5 years ago, I would have said that this was a possibility. But things are changing today. The sharing environment is more protected than you might think. Many new social applications make it easy to set terms-of-use along your sharing path. Any infringement of those terms will be challenged not just by the law, but by your community. Your audience, who benefit form your sharing, can also be the gatekeepers of your rights. Even if you are a traditional copyright holder, this sounds ideal. Furthermore, by realizing all the immediate and emergent rewards that can be had by sharing, you may eventually find that copyright and “All Rights Reserved” are far from your mind. You will enjoy sharing too much to worry about who is keeping a copy. The new economic formula is, the more people remix your works, the higher the return. I want to point out that Sharism is not Communism, nor Socialism. As for those die- hard Communists we know, they have often abused people’s sharing nature and forced them to give up their rights, and their property. Socialism, that tender Communism, in our experience also lacked respect for these rights. Under these systems, the state owns all property. Under Sharism, you can keep ownership, if you want. But I like to share. And this is how I choose to spread ideas, and prosperity Sharism is totally based on your own consensus. It’s not a very hard concept to understand, especially since copyleft movements like the Free Software Foundation and Creative Commons have been around for years. These movements are redefining a more flexible spectrum of licenses for both developers and end-users to tag their works. Because the new licenses can be recognized by either humans or machines, it’s becoming easier to re-share those works in new online ecosystems. The Spirit of the Web, a Social Brain Sharism is the Spirit of the Age of Web 2.0. It has the consistency of a naturalized Epistemology and modernized Axiology, but also promises the power of a new Internet philosophy. Sharism will transform the world into an emergent Social Brain: a networked hybrid of people and software. We are Networked Neurons connected by the synapses of Social Software. This is an evolutionary leap, a small step for us and a giant one for human society. With new “hairy” emergent technologies sprouting all around us, we can generate higher connectivities and increase the throughput of our social links. The more open and strongly connected we social neurons are, the better the sharing environment will be for all people. The more collective our intelligence, the wiser our actions will be. People have always found better solutions through conversations. Now we can put it all online. Sharism will be the politics of the next global superpower. It will not be a country, but a new human network joined by Social Software. This may remain a distant dream, and even a well-defined public sharing policy might not be close at hand. But the ideas that I’m discussing can improve governments today. We can integrate our current and emerging democratic systems with new folksonomies (based on the collaborative, social indexing of information) to enable people to make queries, share data and remix information for public use. The collective intelligence of a vast and equitable sharing environment can be the gatekeeper of our rights, and a government watchdog. In the future, policymaking can be made more nuanced with the micro-involvement of the sharing community. This “Emergent Democracy” is more real-time than periodical parliamentary sessions. It will also increase the spectrum of our choices, beyond the binary options of “Yes” or “No” referenda. Representative democracy will become more timely and diligent, because we will represent ourselves within the system. Sharism will result in better social justice. In a healthy sharing environment, any evidence of injustice can get amplified to get the public’s attention. Anyone who has been abused can get real and instant support from her peers and her peers’ peers. Appeals to justice will take the form of petitions through multiple, interconnected channels. Using these tools, anyone can create a large social impact. With multiple devices and many social applications, each of us can become more sociable, and society more individual. We no longer have to act alone. Emergent democracy will only happen when Sharism becomes the literacy of the majority. Since Sharism can improve communication, collaboration and mutual understanding, I believe it has a place within the educational system. Sharism can be applied to any cultural discourse, CoP (Community of Practice) or problem-solving context. It is also an antidote to social depression, since sharelessness is just dragging our society down. In present or formerly totalitarian countries, this downward cycle is even more apparent. The future world will be a hybrid of human and machine that will generate better and faster decisions anytime, anywhere. The flow of information between minds will become more flexible and more productive. These vast networks of sharing will create a new social order−A Mind Revolution! | A. Form vastly interconnected networks |
What does Fiss mean by Irony?
A. That true freedom of speech calls for the silencing of a few groups.
B. That true freedom of speech calls for the silencing of unorthodox artists, as their work so often offends on a large scale and does not bode positively for the groups the artist hopes to represents.
C. That true freedom of speech depends on the silencing of the state in free speech trials.
D. That true freedom of speech calls for an inspection of the pornography market.
| Shut Up, He Explained Owen Fiss is a professor at the Yale Law School and a highly regarded scholar of constitutional law. The subject of this short book is the present direction of the law governing the freedom of speech. What Professor Fiss has to say about it is worth attending to not merely because of his prominence in the field but because his argument is planted in the common assumptive ground of a lot of contemporary academic thought about the bankruptcy of individualism. The thesis of the book is Fiss', but the wisdom is conventional. Professor Fiss thinks the present direction of First Amendment law is a bad one, and he has an idea about how we might improve it. The short way to put his argument (though it is not quite the way he puts it) is to say that our approach to speech has become increasingly permissive. Courts have become more and more reluctant to allow the state to interfere with the rights of individual speakers to say what they wish, and it is time to roll back that permissiveness and to embark on a new approach that would permit the state to silence some speakers and promote others, but still, Fiss argues, in the name of freedom of speech. This is what Fiss means by the "irony" in his title: that true freedom of speech for all requires suppressing the speech of some. This is not, technically, an irony. It is a paradox. An irony would be the observation that an attempt to increase freedom for all often entails, despite our best efforts, a decrease in freedom for a few. If Fiss had addressed the subject of free speech in this spirit, as an irony, he would undoubtedly have had some interesting things to say, for he is a learned and temperate writer. But he has, instead, chosen to address the issue as an advocate for specific groups he regards as politically disadvantaged--women, gays, victims of racial-hate speech, the poor (or, at least, the not-rich), and people who are critical of market capitalism--and to design a constitutional theory that will enable those groups to enlist the state in efforts either to suppress speech they dislike or to subsidize speech they do like, without running afoul of the First Amendment. Embarked on this task, the most learned and temperate writer in the world would have a hard time avoiding tendentiousness. Fiss does not avoid it. The Irony of Free Speech is a discussion of several speech issues: campaign-finance laws, state funding for the arts, pornography, speech codes, and equal time. These discussions are not doctrinaire, but their general inclination is to favor state intervention, on political grounds, in each of those areas--that is, to favor restrictions on campaign spending, greater regulation of pornography, and so on. Fiss' analyses of specific cases are presented against a lightly sketched historical argument. Light though the sketching is, the historical argument is almost the most objectionable thing about the book, since it involves a distortion of the history of First Amendment law that is fairly plain even to someone who is not a professor at Yale Law School. The argument is that "the liberalism of the nineteenth century was defined by the claims of individual liberty and resulted in an unequivocal demand for liberal government, [while] the liberalism of today embraces the value of equality as well as liberty." The constitutional law of free speech, says Fiss, was shaped by the earlier type of liberalism--he calls it "libertarian"--which regarded free speech as a right of individual self-expression; it is now used to foil efforts to regulate speech in the name of the newer liberal value, equality. Contemporary liberals, inheriting both these traditions, find themselves in a bind. They want, let's say, black students to be free from harassment at institutions where they are, racially, in a minority, since liberals worry that black students cannot be "equal" if they feel intimidated. But those same liberals get upset at the thought of outlawing hate speech, since that would mean infringing upon the right of individuals to express themselves. Fiss' suggestion--this is the chief theoretical proposal of his book--is that liberals should stop thinking about this as a conflict between liberty and equality and start thinking about it as a conflict between two kinds of liberty: social vs. individual. The First Amendment, he says, was intended to foster (in William Brennan's words) "uninhibited, robust, and wide-open" debate in society as a whole; speech that inhibits or monopolizes that debate should therefore fall outside the protection of the law. We can maximize the total freedom of speech by silencing people who prevent others from speaking--when they utter racial epithets, represent women in degrading ways, use their wealth to dominate the press and the political process, or block the funding of unorthodox art. The historical part of this analysis rests on a canard, which is the assertion that the constitutional law of free speech emerged from 19 th -century classical laissez-faire liberalism. It did not. It emerged at the time of World War I, and the principal figures in its creation--Learned Hand, Oliver Wendell Holmes Jr., and Louis Brandeis--were not classical liberals; they were progressives. They abhorred the doctrine of natural rights because, in their time, that doctrine was construed to cover not the right to "self-expression" but the "right to property." Turn-of-the-century courts did not display a libertarian attitude toward civil rights; they displayed a libertarian attitude toward economic rights, tending to throw out legislation aimed at regulating industry and protecting workers on the grounds that people had a constitutional right to enter into contracts and to use their own property as they saw fit. Holmes, Brandeis, and their disciples consistently supported state intervention in economic affairs--the passage of health and safety regulations, the protection of unions, the imposition of taxes, and so on. The post-New Deal liberals whom Fiss associates with the value of equality are their heirs. The heirs of the19 th -century classical liberals are Jack Kemp and Newt Gingrich. Fiss' two "liberalisms" are, in fact, almost entirely different political philosophies. Hand, Holmes, and Brandeis based their First Amendment opinions not on some putative right to individual self-expression (an idea Holmes referred to as "the right of the donkey to drool") but on a democratic need for full and open political debate. First Amendment law since their time has performed its balancing acts on precisely that social value--the very value Fiss now proposes we need to insert into First Amendment jurisprudence. We don't need to insert it, because it was there from the start. Why does Fiss portray the history of First Amendment jurisprudence in this perverted way? Because he wants to line up his own free-speech argument within the conventional academic view that our problems are mostly the consequences of an antiquated and discreditable ideology of liberal individualism, and that they can mostly be solved by adopting a social-constructionist, or communitarian, or "intersubjective" view of human nature instead. The merits of liberal individualism vs. communitarianism can await another occasion to be debated. For since the law governing the freedom of speech does not emerge out of libertarianism, the matter does not boil down to replacing an obsolete belief in "self-expression" with a more up-to-date belief in "robust debate," as Fiss would like to think it does. What it boils down to is whether we need to replace the Hand-Holmes-Brandeis way of maximizing the benefits of free speech in a democratic society, which tries to push the state as far out of the picture as possible, with a different way, which tries to get the state farther into the picture. Here, assuming we want to try the interventionist approach, it is hard to see how a one-size theory can possibly fit all cases. The issues underlying pornography, hate speech, arts grants, campaign finance, and equal-time provisions are all different. The ideological impetus behind judicial developments in the last two areas, campaign finance and equal-time provisions, is related less to speech, except as a kind of constitutional cover, than to a revival of the old "right to property"--that is, the Supreme Court tends to disapprove of legislative and administrative efforts to require broadcasters to carry "opposing viewpoints" on the grounds that since it's their property, owners of television stations should be able to broadcast what they like. Fiss believes that the need for equal-time laws is as urgent today as it was in the 1970s, which is peculiar in light of the proliferation of media outlets. But the state does arguably have an interest, compatible with the First Amendment, in stipulating the way those media are used, and Fiss' discussion of those issues is the least aggravating in his book. Still, that discussion, like his discussions of the other issues, rests on a claim long associated with the left--the claim, in a phrase, that the minority is really the majority. In the case of speech, Fiss appears to believe that the reason the American public is less enlightened than he would wish it to be concerning matters such as feminism, the rights of homosexuals, and regulation of industry is that people are denied access to the opinions and information that would enlighten them. The public is denied this access because the state, in thrall to the ideology of individualism, refuses either to interfere with speech bullies--such as pornographers--who "silence" women, or to subsidize the speech of the unorthodox, such as Robert Mapplethorpe. Fiss' analysis of the Mapplethorpe case offers a good example of the perils of his interventionist approach. Arts policy is, unquestionably, a mess. The solution usually proposed is divorce: Either get the state out of the business altogether or invent some ironclad process for distributing the money using strictly artistic criteria. Fiss rejects both solutions; he wants the criteria to be political. He thinks the NEA should subsidize art that will enhance the "robustness" of the debate and should therefore prefer unorthodox art--though only, of course, if it represents a viewpoint the endowment considers, by virtue of social need and a prior history of exclusion, worthy of its megaphone. (No Nazi art, in other words.) Mapplethorpe's photographs seem to Fiss to qualify under these guidelines, since, he says, "in the late 1980s the AIDS crisis confronted America in the starkest fashion and provoked urgent questions regarding the scope and direction of publicly funded medical research. To address those issues the public--represented by the casual museum visitor--needed an understanding of the lives and practices of the gay community, so long hidden from view." This seems completely wrongheaded. People (for the most part) didn't find Mapplethorpe's X Portfolio photographs objectionable because they depicted homosexuality. They found them objectionable because they depicted sadomasochism. The notion that it was what Fiss calls a "source of empowerment for the members of the gay community" to have homosexuality associated with snarling guys prancing around in leather jockstraps, using bullwhips as sex toys, and pissing in each other's mouths, at a time when AIDS had become a national health problem and the issue of gays in the military was about to arise, is ludicrous. Any NEA chairperson who had the interests of the gay community at heart would have rushed to defund the exhibit. Jesse Helms could not have demonized homosexuality more effectively--which, of course, is why he was pleased to draw public attention to the pictures. Now that is what we call an irony of free speech. Awarding funding to the work of a gay artist because gay Americans need more political clout is an effort at cultural engineering, and the problem with cultural engineering is the problem with social engineering raised to a higher power. We have a hard enough time calculating the effects of the redistribution of wealth in our society. How can we possibly calculate the effects of redistributing the right to speak--of taking it away from people Professor Fiss feels have spoken long enough and mandating it for people he feels have not been adequately heard? One thing that is plain from the brief unhappy history of campus speech codes is that you automatically raise the value of the speech you punish and depress the value of the speech you sponsor. There are indeed many ironies here. Maybe someone will write a book about them. | A. That true freedom of speech calls for the silencing of a few groups. |
Why does the Captain go looking for Purnie?
A. The Captain knows that an animal with Purnie's strength is worth a fortune.
B. The Captian knows an animal that can stop time is worth a fortune.
C. The Captain knows a radioactive animal is worth a fortune.
D. The Captain knows Purnie saved the crew.
| BEACH SCENE By MARSHALL KING Illustrated by WOOD [Transcriber's Note: This etext was produced from Galaxy Magazine October 1960. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed.] It was a fine day at the beach for Purnie's game—but his new friends played very rough! Purnie ran laughing and shouting through the forest until he could run no more. He fell headlong into a patch of blue moss and whooped with delight in having this day free for exploring. He was free to see the ocean at last. When he had caught his breath, he looked back through the forest. No sign of the village; he had left it far behind. Safe from the scrutiny of brothers and parents, there was nothing now to stop him from going to the ocean. This was the moment to stop time. "On your mark!" he shouted to the rippling stream and its orange whirlpools. He glanced furtively from side to side, pretending that some object might try to get a head start. "Get set!" he challenged the thin-winged bees that hovered over the abundant foliage. "Stop!" He shrieked this command upward toward the dense, low-hanging purple clouds that perennially raced across the treetops, making one wonder how tall the trees really were. His eyes took quick inventory. It was exactly as he knew it would be: the milky-orange stream had become motionless and its minute whirlpools had stopped whirling; a nearby bee hung suspended over a paka plant, its transparent wings frozen in position for a downward stroke; and the heavy purple fluid overhead held fast in its manufacture of whorls and nimbi. With everything around him in a state of perfect tableau, Purnie hurried toward the ocean. If only the days weren't so short! he thought. There was so much to see and so little time. It seemed that everyone except him had seen the wonders of the beach country. The stories he had heard from his brothers and their friends had taunted him for as long as he could remember. So many times had he heard these thrilling tales that now, as he ran along, he could clearly picture the wonderland as though he were already there. There would be a rockslide of petrified logs to play on, the ocean itself with waves higher than a house, the comical three-legged tripons who never stopped munching on seaweed, and many kinds of other wonderful creatures found only at the ocean. He bounced through the forest as though the world was reserved this day just for him. And who could say it wasn't? he thought. Wasn't this his fifth birthday? He ran along feeling sorry for four-year-olds, and even for those who were only four and a half, for they were babies and wouldn't dare try slipping away to the ocean alone. But five! "I'll set you free, Mr. Bee—just wait and see!" As he passed one of the many motionless pollen-gathering insects he met on the way, he took care not to brush against it or disturb its interrupted task. When Purnie had stopped time, the bees—like all the other creatures he met—had been arrested in their native activities, and he knew that as soon as he resumed time, everything would pick up where it had left off. When he smelled an acid sweetness that told him the ocean was not far off, his pulse quickened in anticipation. Rather than spoil what was clearly going to be a perfect day, he chose to ignore the fact that he had been forbidden to use time-stopping as a convenience for journeying far from home. He chose to ignore the oft-repeated statement that an hour of time-stopping consumed more energy than a week of foot-racing. He chose to ignore the negative maxim that "small children who stop time without an adult being present, may not live to regret it." He chose, instead, to picture the beaming praise of family and friends when they learned of his brave journey. The journey was long, the clock stood still. He stopped long enough to gather some fruit that grew along the path. It would serve as his lunch during this day of promise. With it under his arm he bounded along a dozen more steps, then stopped abruptly in his tracks. He found himself atop a rocky knoll, overlooking the mighty sea! He was so overpowered by the vista before him that his "Hurrah!" came out as a weak squeak. The ocean lay at the ready, its stilled waves awaiting his command to resume their tidal sweep. The breakers along the shoreline hung in varying stages of disarray, some having already exploded into towering white spray while others were poised in smooth orange curls waiting to start that action. And there were new friends everywhere! Overhead, a flock of spora were frozen in a steep glide, preparatory to a beach landing. Purnie had heard of these playful creatures many times. Today, with his brothers in school, he would have the pets all to himself. Further down the beach was a pair of two-legged animals poised in mid-step, facing the spot where Purnie now stood. Some distance behind them were eight more, each of whom were motionless in a curious pose of interrupted animation. And down in the water, where the ocean ran itself into thin nothingness upon the sand, he saw standing here and there the comical tripons, those three-legged marine buffoons who made handsome careers of munching seaweed. "Hi there!" Purnie called. When he got no reaction, he remembered that he himself was "dead" to the living world: he was still in a zone of time-stopping, on the inside looking out. For him, the world would continue to be a tableau of mannikins until he resumed time. "Hi there!" he called again; but now his mental attitude was that he expected time to resume. It did! Immediately he was surrounded by activity. He heard the roar of the crashing orange breakers, he tasted the dew of acid that floated from the spray, and he saw his new friends continue the actions which he had stopped while back in the forest. He knew, too, that at this moment, in the forest, the little brook picked up its flow where it had left off, the purple clouds resumed their leeward journey up the valley, and the bees continued their pollen-gathering without having missed a single stroke of their delicate wings. The brook, the clouds, and the insects had not been interrupted in the least; their respective tasks had been performed with continuing sureness. It was time itself that Purnie had stopped, not the world around him. He scampered around the rockpile and down the sandy cliff to meet the tripons who, to him, had just come to life. "I can stand on my head!" He set down his lunch and balanced himself bottoms-up while his legs pawed the air in an effort to hold him in position. He knew it was probably the worst head-stand he had ever done, for he felt weak and dizzy. Already time-stopping had left its mark on his strength. But his spirits ran on unchecked. The tripon thought Purnie's feat was superb. It stopped munching long enough to give him a salutory wag of its rump before returning to its repast. Purnie ran from pillar to post, trying to see and do everything at once. He looked around to greet the flock of spora, but they had glided to a spot further along the shore. Then, bouncing up to the first of the two-legged animals, he started to burst forth with his habitual "Hi there!" when he heard them making sounds of their own. "... will be no limit to my operations now, Benson. This planet makes seventeen. Seventeen planets I can claim as my own!" "My, my. Seventeen planets. And tell me, Forbes, just what the hell are you going to do with them—mount them on the wall of your den back in San Diego?" "Hi there, wanna play?" Purnie's invitation got nothing more than startled glance from the animals who quickly returned to their chatter. He scampered up the beach, picked up his lunch, and ran back to them, tagging along at their heels. "I've got my lunch, want some?" "Benson, you'd better tell your men back there to stop gawking at the scenery and get to work. Time is money. I didn't pay for this expedition just to give your flunkies a vacation." The animals stopped so suddenly that Purnie nearly tangled himself in their heels. "All right, Forbes, just hold it a minute. Listen to me. Sure, it's your money that put us here; it's your expedition all the way. But you hired me to get you here with the best crew on earth, and that's just what I've done. My job isn't over yet. I'm responsible for the safety of the men while we're here, and for the safe trip home." "Precisely. And since you're responsible, get 'em working. Tell 'em to bring along the flag. Look at the damn fools back there, playing in the ocean with a three-legged ostrich!" "Good God, man, aren't you human? We've only been on this planet twenty minutes! Naturally they want to look around. They half expected to find wild animals or worse, and here we are surrounded by quaint little creatures that run up to us like we're long-lost brothers. Let the men look around a minute or two before we stake out your claim." "Bah! Bunch of damn children." As Purnie followed along, a leg shot out at him and missed. "Benson, will you get this bug-eyed kangaroo away from me!" Purnie shrieked with joy at this new frolic and promptly stood on his head. In this position he got an upside down view of them walking away. He gave up trying to stay with them. Why did they move so fast, anyway? What was the hurry? As he sat down and began eating his lunch, three more of the creatures came along making excited noises, apparently trying to catch up to the first two. As they passed him, he held out his lunch. "Want some?" No response. Playing held more promise than eating. He left his lunch half eaten and went down to where they had stopped further along the beach. "Captain Benson, sir! Miles has detected strong radiation in the vicinity. He's trying to locate it now." "There you are, Forbes. Your new piece of real estate is going to make you so rich that you can buy your next planet. That'll make eighteen, I believe." "Radiation, bah! We've found low-grade ore on every planet I've discovered so far, and this one'll be no different. Now how about that flag? Let's get it up, Benson. And the cornerstone, and the plaque." "All right, lads. The sooner we get Mr. Forbes's pennant raised and his claim staked out, the sooner we can take time to look around. Lively now!" When the three animals went back to join the rest of their group, the first two resumed walking. Purnie followed along. "Well, Benson, you won't have to look far for materials to use for the base of the flag pole. Look at that rockpile up there. "Can't use them. They're petrified logs. The ones on top are too high to carry down, and if we move those on the bottom, the whole works will slide down on top of us." "Well—that's your problem. Just remember, I want this flag pole to be solid. It's got to stand at least—" "Don't worry, Forbes, we'll get your monument erected. What's this with the flag? There must be more to staking a claim than just putting up a flag." "There is, there is. Much more. I've taken care of all requirements set down by law to make my claim. But the flag? Well, you might say it represents an empire, Benson. The Forbes Empire. On each of my flags is the word FORBES, a symbol of development and progress. Call it sentiment if you will." "Don't worry, I won't. I've seen real-estate flags before." "Damn it all, will you stop referring to this as a real-estate deal? What I'm doing is big, man. Big! This is pioneering." "Of course. And if I'm not mistaken, you've set up a neat little escrow system so that you not only own the planets, but you will virtually own the people who are foolish enough to buy land on them." "I could have your hide for talking to me like this. Damn you, man! It's people like me who pay your way. It's people like me who give your space ships some place to go. It's people like me who pour good money into a chancey job like this, so that people like you can get away from thirteen-story tenement houses. Did you ever think of that?" "I imagine you'll triple your money in six months." When they stopped, Purnie stopped. At first he had been interested in the strange sounds they were making, but as he grew used to them, and as they in turn ignored his presence, he hopped alongside chattering to himself, content to be in their company. He heard more of these sounds coming from behind, and he turned to see the remainder of the group running toward them. "Captain Benson! Here's the flag, sir. And here's Miles with the scintillometer. He says the radiation's getting stronger over this way!" "How about that, Miles?" "This thing's going wild, Captain. It's almost off scale." Purnie saw one of the animals hovering around him with a little box. Thankful for the attention, he stood on his head. "Can you do this?" He was overjoyed at the reaction. They all started making wonderful noises, and he felt most satisfied. "Stand back, Captain! Here's the source right here! This little chuck-walla's hotter than a plutonium pile!" "Let me see that, Miles. Well, I'll be damned! Now what do you suppose—" By now they had formed a widening circle around him, and he was hard put to think of an encore. He gambled on trying a brand new trick: he stood on one leg. "Benson, I must have that animal! Put him in a box." "Now wait a minute, Forbes. Universal Law forbids—" "This is my planet and I am the law. Put him in a box!" "With my crew as witness, I officially protest—" "Good God, what a specimen to take back. Radio-active animals! Why, they can reproduce themselves, of course! There must be thousands of these creatures around here someplace. And to think of those damn fools on Earth with their plutonium piles! Hah! Now I'll have investors flocking to me. How about it, Benson—does pioneering pay off or doesn't it?" "Not so fast. Since this little fellow is radioactive, there may be great danger to the crew—" "Now look here! You had planned to put mineral specimens in a lead box, so what's the difference? Put him in a box." "He'll die." "I have you under contract, Benson! You are responsible to me, and what's more, you are on my property. Put him in a box." Purnie was tired. First the time-stopping, then this. While this day had brought more fun and excitement than he could have hoped for, the strain was beginning to tell. He lay in the center of the circle happily exhausted, hoping that his friends would show him some of their own tricks. He didn't have to wait long. The animals forming the circle stepped back and made way for two others who came through carrying a box. Purnie sat up to watch the show. "Hell, Captain, why don't I just pick him up? Looks like he has no intention of running away." "Better not, Cabot. Even though you're shielded, no telling what powers the little fella has. Play it safe and use the rope." "I swear he knows what we're saying. Look at those eyes." "All right, careful now with that line." "Come on, baby. Here you go. That's a boy!" Purnie took in these sounds with perplexed concern. He sensed the imploring quality of the creature with the rope, but he didn't know what he was supposed to do. He cocked his head to one side as he wiggled in anticipation. He saw the noose spinning down toward his head, and, before he knew it, he had scooted out of the circle and up the sandy beach. He was surprised at himself for running away. Why had he done it? He wondered. Never before had he felt this fleeting twinge that made him want to protect himself. He watched the animals huddle around the box on the beach, their attention apparently diverted to something else. He wished now that he had not run away; he felt he had lost his chance to join in their fun. "Wait!" He ran over to his half-eaten lunch, picked it up, and ran back into the little crowd. "I've got my lunch, want some?" The party came to life once more. His friends ran this way and that, and at last Purnie knew that the idea was to get him into the box. He picked up the spirit of the tease, and deliberately ran within a few feet of the lead box, then, just as the nearest pursuer was about to push him in, he sidestepped onto safer ground. Then he heard a deafening roar and felt a warm, wet sting in one of his legs. "Forbes, you fool! Put away that gun!" "There you are, boys. It's all in knowing how. Just winged him, that's all. Now pick him up." The pang in his leg was nothing: Purnie's misery lay in his confusion. What had he done wrong? When he saw the noose spinning toward him again, he involuntarily stopped time. He knew better than to use this power carelessly, but his action now was reflex. In that split second following the sharp sting in his leg, his mind had grasped in all directions to find an acceptable course of action. Finding none, it had ordered the stoppage of time. The scene around him became a tableau once more. The noose hung motionless over his head while the rest of the rope snaked its way in transverse waves back to one of the two-legged animals. Purnie dragged himself through the congregation, whimpering from his inability to understand. As he worked his way past one creature after another, he tried at first to not look them in the eye, for he felt sure he had done something wrong. Then he thought that by sneaking a glance at them as he passed, he might see a sign pointing to their purpose. He limped by one who had in his hand a small shiny object that had been emitting smoke from one end; the smoke now billowed in lifeless curls about the animal's head. He hobbled by another who held a small box that had previously made a hissing sound whenever Purnie was near. These things told him nothing. Before starting his climb up the knoll, he passed a tripon which, true to its reputation, was comical even in fright. Startled by the loud explosion, it had jumped four feet into the air before Purnie had stopped time. Now it hung there, its beak stuffed with seaweed and its three legs drawn up into a squatting position. Leaving the assorted statues behind, he limped his way up the knoll, torn between leaving and staying. What an odd place, this ocean country! He wondered why he had not heard more detail about the beach animals. Reaching the top of the bluff, he looked down upon his silent friends with a feeling of deep sorrow. How he wished he were down there playing with them. But he knew at last that theirs was a game he didn't fit into. Now there was nothing left but to resume time and start the long walk home. Even though the short day was nearly over, he knew he didn't dare use time-stopping to get himself home in nothing flat. His fatigued body and clouded mind were strong signals that he had already abused this faculty. When Purnie started time again, the animal with the noose stood in open-mouthed disbelief as the rope fell harmlessly to the sand—on the spot where Purnie had been standing. "My God, he's—he's gone." Then another of the animals, the one with the smoking thing in his hand, ran a few steps toward the noose, stopped and gaped at the rope. "All right, you people, what's going on here? Get him in that box. What did you do with him?" The resumption of time meant nothing at all to those on the beach, for to them time had never stopped. The only thing they could be sure of was that at one moment there had been a fuzzy creature hopping around in front of them, and the next moment he was gone. "Is he invisible, Captain? Where is he?" "Up there, Captain! On those rocks. Isn't that him?" "Well, I'll be damned!" "Benson, I'm holding you personally responsible for this! Now that you've botched it up, I'll bring him down my own way." "Just a minute, Forbes, let me think. There's something about that fuzzy little devil that we should.... Forbes! I warned you about that gun!" Purnie moved across the top of the rockpile for a last look at his friends. His weight on the end of the first log started the slide. Slowly at first, the giant pencils began cascading down the short distance to the sand. Purnie fell back onto solid ground, horrified at the spectacle before him. The agonizing screams of the animals below filled him with hysteria. The boulders caught most of them as they stood ankle-deep in the surf. Others were pinned down on the sand. "I didn't mean it!" Purnie screamed. "I'm sorry! Can't you hear?" He hopped back and forth near the edge of the rise, torn with panic and shame. "Get up! Please get up!" He was horrified by the moans reaching his ears from the beach. "You're getting all wet! Did you hear me? Please get up." He was choked with rage and sorrow. How could he have done this? He wanted his friends to get up and shake themselves off, tell him it was all right. But it was beyond his power to bring it about. The lapping tide threatened to cover those in the orange surf. Purnie worked his way down the hill, imploring them to save themselves. The sounds they made carried a new tone, a desperate foreboding of death. "Rhodes! Cabot! Can you hear me?" "I—I can't move, Captain. My leg, it's.... My God, we're going to drown!" "Look around you, Cabot. Can you see anyone moving?" "The men on the beach are nearly buried, Captain. And the rest of us here in the water—" "Forbes. Can you see Forbes? Maybe he's—" His sounds were cut off by a wavelet gently rolling over his head. Purnie could wait no longer. The tides were all but covering one of the animals, and soon the others would be in the same plight. Disregarding the consequences, he ordered time to stop. Wading down into the surf, he worked a log off one victim, then he tugged the animal up to the sand. Through blinding tears, Purnie worked slowly and carefully. He knew there was no hurry—at least, not as far as his friends' safety was concerned. No matter what their condition of life or death was at this moment, it would stay the same way until he started time again. He made his way deeper into the orange liquid, where a raised hand signalled the location of a submerged body. The hand was clutching a large white banner that was tangled among the logs. Purnie worked the animal free and pulled it ashore. It was the one who had been carrying the shiny object that spit smoke. Scarcely noticing his own injured leg, he ferried one victim after another until there were no more in the surf. Up on the beach, he started unraveling the logs that pinned down the animals caught there. He removed a log from the lap of one, who then remained in a sitting position, his face contorted into a frozen mask of agony and shock. Another, with the weight removed, rolled over like an iron statue into a new position. Purnie whimpered in black misery as he surveyed the chaotic scene before him. At last he could do no more; he felt consciousness slipping away from him. He instinctively knew that if he lost his senses during a period of time-stopping, events would pick up where they had left off ... without him. For Purnie, this would be death. If he had to lose consciousness, he knew he must first resume time. Step by step he plodded up the little hill, pausing every now and then to consider if this were the moment to start time before it was too late. With his energy fast draining away, he reached the top of the knoll, and he turned to look down once more on the group below. Then he knew how much his mind and body had suffered: when he ordered time to resume, nothing happened. His heart sank. He wasn't afraid of death, and he knew that if he died the oceans would roll again and his friends would move about. But he wanted to see them safe. He tried to clear his mind for supreme effort. There was no urging time to start. He knew he couldn't persuade it by bits and pieces, first slowly then full ahead. Time either progressed or it didn't. He had to take one viewpoint or the other. Then, without knowing exactly when it happened, his mind took command.... His friends came to life. The first one he saw stir lay on his stomach and pounded his fists on the beach. A flood of relief settled over Purnie as sounds came from the animal. "What's the matter with me? Somebody tell me! Am I nuts? Miles! Schick! What's happening?" "I'm coming, Rhodes! Heaven help us, man—I saw it, too. We're either crazy or those damn logs are alive!" "It's not the logs. How about us? How'd we get out of the water? Miles, we're both cracking." "I'm telling you, man, it's the logs, or rocks or whatever they are. I was looking right at them. First they're on top of me, then they're piled up over there!" "Damnit, the logs didn't pick us up out of the ocean, did they? Captain Benson!" "Are you men all right?" "Yes sir, but—" "Who saw exactly what happened?" "I'm afraid we're not seeing right, Captain. Those logs—" "I know, I know. Now get hold of yourselves. We've got to round up the others and get out of here while time is on our side." "But what happened, Captain?" "Hell, Rhodes, don't you think I'd like to know? Those logs are so old they're petrified. The whole bunch of us couldn't lift one. It would take super-human energy to move one of those things." "I haven't seen anything super-human. Those ostriches down there are so busy eating seaweed—" "All right, let's bear a hand here with the others. Some of them can't walk. Where's Forbes?" "He's sitting down there in the water, Captain, crying like a baby. Or laughing. I can't tell which." "We'll have to get him. Miles, Schick, come along. Forbes! You all right?" "Ho-ho-ho! Seventeen! Seventeen! Seventeen planets, Benson, and they'll do anything I say! This one's got a mind of its own. Did you see that little trick with the rocks? Ho-ho!" "See if you can find his gun, Schick; he'll either kill himself or one of us. Tie his hands and take him back to the ship. We'll be along shortly." "Hah-hah-hah! Seventeen! Benson, I'm holding you personally responsible for this. Hee-hee!" Purnie opened his eyes as consciousness returned. Had his friends gone? He pulled himself along on his stomach to a position between two rocks, where he could see without being seen. By the light of the twin moons he saw that they were leaving, marching away in groups of two and three, the weak helping the weaker. As they disappeared around the curving shoreline, the voices of the last two, bringing up the rear far behind the others, fell faintly on his ears over the sound of the surf. "Is it possible that we're all crazy, Captain?" "It's possible, but we're not." "I wish I could be sure." "See Forbes up ahead there? What do you think of him?" "I still can't believe it." "He'll never be the same." "Tell me something. What was the most unusual thing you noticed back there?" "You must be kidding, sir. Why, the way those logs were off of us suddenly—" "Yes, of course. But I mean beside that." "Well, I guess I was kind of busy. You know, scared and mixed up." "But didn't you notice our little pop-eyed friend?" "Oh, him. I'm afraid not, Captain. I—I guess I was thinking mostly of myself." "Hmmm. If I could only be sure I saw him. If only someone else saw him too." "I'm afraid I don't follow you, sir." "Well, damn it all, you know that Forbes took a pot shot at him. Got him in the leg. That being the case, why would the fuzzy little devil come back to his tormentors—back to us—when we were trapped under those logs?" "Well, I guess as long as we were trapped, he figured we couldn't do him any more harm.... I'm sorry, that was a stupid answer. I guess I'm still a little shaky." "Forget it. Look, you go ahead to the ship and make ready for take-off. I'll join you in a few minutes. I think I'll go back and look around. You know. Make sure we haven't left anyone." "No need to do that. They're all ahead of us. I've checked." "That's my responsibility, Cabot, not yours. Now go on." As Purnie lay gathering strength for the long trek home, he saw through glazed eyes one of the animals coming back along the beach. When it was nearly directly below him, he could hear it making sounds that by now had become familiar. "Where are you?" Purnie paid little attention to the antics of his friend; he was beyond understanding. He wondered what they would say at home when he returned. "We've made a terrible mistake. We—" The sounds faded in and out on Purnie's ears as the creature turned slowly and called in different directions. He watched the animal walk over to the pile of scattered logs and peer around and under them. "If you're hurt I'd like to help!" The twin moons were high in the sky now, and where their light broke through the swirling clouds a double shadow was cast around the animal. With foggy awareness, Purnie watched the creature shake its head slowly, then walk away in the direction of the others. Purnie's eyes stared, without seeing, at the panorama before him. The beach was deserted now, and his gaze was transfixed on a shimmering white square floating on the ocean. Across it, the last thing Purnie ever saw, was emblazoned the word FORBES. | D. The Captain knows Purnie saved the crew. |
Which isn't true of the test results?
A. some people were able to identify the beer based on taste
B. not all people knew beers as well as they thought they did
C. American beers typically scored higher
D. Hefeweizens were not popular among the testers
| More Booze You Can Use When we last heard from them, the members of the Slate beer-testing team were coping with lagers and trying to see if they could taste the 3-to-1 price difference between the most- and least-expensive brands. (Click for a wrap-up of the first round of beer tasting.) The answer was: They found one beer they really liked, Samuel Adams Boston Lager , and one they really hated, imported Grolsch from Holland. Both were expensive beers--Grolsch was the most expensive in the test--and otherwise the testers had a hard time telling beers apart. The members of the team, as noted in the original article, all hold day jobs at Microsoft, mainly as designers, managers, and coders for Microsoft Word. The point of the second test was not to find the difference between cheap and expensive beers but instead to compare a variety of top-of-the-line beers. Was there one kind the tasters preferred consistently? Could they detect any of the subtleties of brewing style and provenance that microbrew customers pay such attention to when choosing some Doppelbock over a cream ale? Since the tasting panel had left the first round grumbling that cheap lagers were not a fair test of their abilities, this second round of testing was advertised to the panel as a reward. Every beer in Round 2 would be a fancy beer. A microbrew. A "craft beer." A prestigious import. These were the kinds of beer the panel members said they liked--and the ones they said they were most familiar with. One aspect of the reward was that they would presumably enjoy the actual testing more--fewer rueful beer descriptions along the lines of "urine" or "get it away!" were expected than in the first round. The other aspect of anticipated reward was the panelists' unspoken but obvious assumption that this time they would "do better" on the test. Intellectual vanity being what it is, people who had fought for and won jobs at Microsoft and who still must fight every six months for primacy on the employee-ranking scale (which determines--gasp!--how many new stock options they receive) would assume that their skill as tasters was on trial, just as much as the beer was. Of course they were right, which is what made this round as amusing to administer as the first one had been. Here is what happened and what it meant: 1. Procedure. This was similar in most ways to the experimental approach of Round 1. The nine testers who showed up were a subset of the original 12. The missing three dropped out with excuses of "my wife is sick" (one person) and "meeting is running long" (two). As before, each tester found before him on a table 10 red plastic cups, labeled A through J. Each cup held 3 ounces of one of the beers. The A-to-J labeling scheme was the same for all testers. Instead of saltines for palate-cleansing, this time we had popcorn and nuts. As they began, the tasters were given these and only these clues: that the flight included one "holdover" beer from the previous round (Sam Adams); that it included at least one import (Bass); that it included at least one macrobrew , specifically, a member of the vast Anheuser-Busch family (Michelob Hefeweizen). After sampling all beers, the tasters rated them as follows: Overall quality points, from zero to 100, reflecting their personal, subjective fondness for the beer. Descriptions of and comments about each beer's taste--"smooth and nutty," "too strong," etc. If the first ranking was a measure of how good each beer was, this was an attempt to explain what made it good. Best and Worst , one of each from the group. Name that beer! The tasters were told that some of the drinks were Hefeweizens, some might be IPAs (India pale ales), some might be bitters, and so on. They were asked to put each beer in its proper category--and to name a specific brewery and brand if they could. The idea here was to test the veteran beer drinkers' claim to recognize the distinctive tastes of famous brands. (To see all the grids for all the beers, click .) 2. Philosophy. The first round of testing was All Lager. This second round was All Fancy, and Mainly Not Lager. As several correspondents (for instance, the of Best American Beers ) have helpfully pointed out, the definition of lager provided last time was not exactly "accurate." If you want to stay within the realm of textbook definitions, a lager is a beer brewed a particular way--slowly, at cool temperatures, with yeast that settles on the bottom of the vat. This is in contrast with an ale, which is brewed faster, warmer, and with the yeast on top. By this same reasoning, lagers don't have to be light-colored, weak-flavored, and watery, as mainstream American lagers are. In principle, lagers can be dark, fierce, manly. Therefore, the correspondents suggest, it was wrong to impugn Sam Adams or Pete's Wicked for deceptive labeling, in presenting their tawnier, more flavorful beers as lagers too. To this the beer scientist must say: Book-learning is fine in its place. But let's be realistic. Actual drinking experience teaches the American beer consumer that a) all cheap beers are lagers; and b) most lagers are light-colored and weak. The first test was designed to evaluate low-end beers and therefore had to be lager-centric. This one is designed to test fancy beers--but in the spirit of open-mindedness and technical accuracy, it includes a few "strong" lagers too. 3. Materials. The 10 test beers were chosen with several goals in mind: To cover at least a modest range of fancy beer types--extra special bitter, India pale ale, Hefeweizen, and so on. To include both imported and domestic beers. Among the domestic microbrews, there's an obvious skew toward beers from the Pacific Northwest. But as Microsoft would put it, that's a feature not a bug. These beers all came from the Safeway nearest the Redmond, Wash., "main campus" of Microsoft, and microbrews are supposed to be local. To include one holdover from the previous test, as a scientific control on our tasters' preferences. This was Sam Adams , runaway winner of Round 1. To include one fancy product from a monster-scale U.S. mass brewery, to see if the tasters liked it better or worse than the cute little microbrews. This was Michelob Hefeweizen , from the pride of St. Louis, Anheuser-Busch. Click for pricing information and pre-quaffing evaluations. The beers tasted were: 4. Data Analysis. a) Best and Worst. Compared to the lager test, we would expect the range of "best" choices to be more varied, since all the tested beers were supposed to be good. This expectation was most dramatically borne out in the "Best and Worst" rankings. The nine tasters cast a total of nine Worst votes and 11.5 Best votes. (Tester No. 1 turned in a sheet with three Best selections, or two more than his theoretical quota. Tester No. 4 listed a Best and a Best-minus, which counted as half a vote.) The results were clearest at the bottom: three Worsts for Pyramid Hefeweizen , even though most comments about the beer were more or less respectful. ("Bitter, drinkable.") But at the top and middle the situation was muddier: There were three Bests for Full Sail ESB , which most of the tasters later said they weren't familiar with, and 2.5 for Redhook IPA , which all the tasters knew. But each of these also got a Worst vote, and most of the other beers had a mixed reading. So far, the tasters are meeting expectations, finding something to like in nearly all these fancy beers. b) Overall preference points. Here the complications increase. The loser was again apparent: Pyramid Hefeweizen came in last on rating points, as it had in the Best/Worst derby. But the amazing dark horse winner was Michelob Hefeweizen . The three elements of surprise here, in ascending order of unexpectedness, are: This best-liked beer belonged to the same category, Hefeweizen, as the least-liked product, from Pyramid. This was also the only outright Anheuser-Busch product in the contest (the Redhooks are 75 percent A-B free). It is safe to say that all tasters would have said beforehand that they would rank an American macrobrew last, and Anheuser-Busch last of all. Although it clearly won on overall preference points, Michelob Hefeweizen was the only beer not to have received a single "Best" vote. The first two anomalies can be written off as testament to the power of a blind taste test. The third suggests an important difference in concepts of "bestness." Sometimes a product seems to be the best of a group simply because it's the most unusual or distinctive. This is why very high Wine Spectator ratings often go to wines that mainly taste odd. But another kind of bestness involves an unobtrusive, day-in day-out acceptability. That seems to be Michelob Hefe 's achievement here: no one's first choice, but high on everyone's list. Let's go to the charts: This table shows how the beers performed on "raw score"--that is, without the advanced statistical adjustment of throwing out the highest and lowest score each beer received. Next, we have "corrected average preference points," throwing out the high and low marks for each beer. The result is basically the same: It is worth noting the fate of Sam Adams on these charts. Here it ends up with a score of less than 61. These were the numbers awarded by the very same tasters who gave it a corrected preference rating of 83.33 the last time around--and 10 "Best" votes, vs. one Best (and one Worst) this time. The shift in Bests is understandable and demonstrates the importance of picking your competition. The severe drop in preference points illustrates more acutely the ancient principle of being a big fish in a small pond. These same tasters thought that Sam Adams was objectively much better when it was surrounded by Busch and Schmidt's. c) Value rankings. Last time this calculation led to what the colorful French would call a bouleversement. One of the cheapest beers, Busch, which had been in the lower ranks on overall preference points, came out at the top on value-for-money ratings, because it was so cheap. The big surprise now is that the highest-rated beer was also the cheapest one, Michelob Hefe , so the value calculation turned into a rout: Pyramid Hefeweizen was expensive on top of being unpopular, so its position at the bottom was hammered home--but not as painfully as that of Bass Ale . Bass had been in the respectable lower middle class of the preference rankings, so its disappointing Val-u-meter showing mainly reflects the fact that it was the only beer not on "sale" and therefore by far the costliest entry in the experiment. d) Taster skill. As members of the tasting panel began to suspect, they themselves were being judged while they judged the beer. One of the tasters, No. 7, decided to live dangerously and give specific brands and breweries for Samples A through J. This man was the only panel member whose job does not involve designing Microsoft Word--and the only one to identify two or more of the beers accurately and specifically. (He spotted Redhook IPA and Redhook ESB.) The fact that the beers correctly identified were the two most popular microbrews in the Seattle area suggests that familiarity is the main ingredient in knowing your beer. Many others were simply lost. Barely half the tasters, five of nine, recognized that Michelob Hefeweizen was a Hefeweizen. Before the test, nine of nine would have said that picking out a Hefe was easy, because of its cloudy look and wheaty flavor. Three tasters thought Sam Adams was an IPA ; two thought Redhook's IPA was a Hefeweizen. In fairness, six of nine testers identified Pyramid Hefeweizen as a Hefe, and six recognized Full Sail ESB as a bitter. Much in the fashion of blind men describing an elephant, here is a how the testers handled Sam Adams Boston Lager : 5. Implications and Directions for Future Research. Science does not always answer questions; often, it raises many new ones. This excursion into beer science mainly raises the question: What kind of people are we? If we are Gradgrind-like empiricists, living our life for "welfare maximization" as described in introductory econ. courses, the conclusion is obvious. We learned from the first experiment to buy either Sam Adams (when we wanted maximum lager enjoyment per bottle) or Busch (for maximum taste and snob appeal per dollar). From this second round we see an even more efficient possibility: Buy Michelob Hefeweizen and nothing else, since on the basis of this test it's the best liked and the cheapest beer. By the way, if there is a single company whose achievements the testing panel honored, it would be Anheuser-Busch . From its brewing tanks came two of the double-crown winners of the taste tests: plain old Busch , the Taste-o-meter and Snob-o-meter victor of Round 1, and Michelob Hefeweizen , the preference-point and Val-u-meter winner this time. But, of course, there is another possibility: that what is excluded in a blind taste test is in fact what we want, and are happy to pay for, when we sit down with a beer. The complicated label, the fancy bottle, the exotic concept that this beer has traveled from some far-off corner of Bohemia or even the Yakima Valley--all this may be cheap at the $1.25-per-pint cost difference between the cheapest and the most expensive beers. In elementary school, we all endured a standard science experiment: If you shut your eyes and pinch your nose closed, can you tell any difference in the taste of a slice of apple, of carrot, of pear? You can't--but that doesn't mean that from then on you should close your eyes, hold your nose, and chew a cheap carrot when you feel like having some fruit. There is a time and place for carrots, but also for juicy pears. There is a time for Busch, but also for Full Sail "Equinox." For scientists who want to continue this work at home, here are a few suggestions for further research: Tell the testers ahead of time what beers they will be drinking. Ask them to rank the beers, 1 through 10, based on how well they like them. Then compare the list with the "revealed preferences" that come from the blind test. As a variation, show them the list ahead of time and ask them to pick out the beer they know they love and the one they know they hate. Then compare this with the "after" list. If you're going to test imported lagers, try Foster's or Corona rather than Grolsch. Remember to stay strictly in the scientist's role. Don't take the test yourself. | D. Hefeweizens were not popular among the testers |
Which tasks do they apply their method to? | ### Introduction
Author contributions: Hao Zhu designed the research; Weize Chen prepared the data, and organized data annotation; Hao Zhu and Xu Han designed the experiments; Weize Chen performed the experiments; Hao Zhu, Weize Chen and Xu Han wrote the paper; Zhiyuan Liu and Maosong Sun proofread the paper. Zhiyuan Liu is the corresponding author. Relations, representing various types of connections between entities or arguments, are the core of expressing relational facts in most general knowledge bases (KBs) BIBREF0 , BIBREF1 . Hence, identifying relations is a crucial problem for several information extraction tasks. Although considerable effort has been devoted to these tasks, some nuances between similar relations are still overlooked, (tab:similarityexample shows an example); on the other hand, some distinct surface forms carrying the same relational semantics are mistaken as different relations. These severe problems motivate us to quantify the similarity between relations in a more effective and robust method. In this paper, we introduce an adaptive and general framework for measuring similarity of the pairs of relations. Suppose for each relation INLINEFORM0 , we have obtained a conditional distribution, INLINEFORM1 ( INLINEFORM2 are head and tail entities, and INLINEFORM3 is a relation), over all head-tail entity pairs given INLINEFORM4 . We could quantify similarity between a pair of relations by the divergence between the conditional probability distributions given these relations. In this paper, this conditional probability is given by a simple feed-forward neural network, which can capture the dependencies between entities conditioned on specific relations. Despite its simplicity, the proposed network is expected to cover various facts, even if the facts are not used for training, owing to the good generalizability of neural networks. For example, our network will assign a fact a higher probability if it is “logical”: e.g., the network might prefer an athlete has the same nationality as same as his/her national team rather than other nations. Intuitively, two similar relations should have similar conditional distributions over head-tail entity pairs INLINEFORM0 , e.g., the entity pairs associated with be trade to and play for are most likely to be athletes and their clubs, whereas those associated with live in are often people and locations. In this paper, we evaluate the similarity between relations based on their conditional distributions over entity pairs. Specifically, we adopt Kullback–Leibler (KL) divergence of both directions as the metric. However, computing exact KL requires iterating over the whole entity pair space INLINEFORM1 , which is quite intractable. Therefore, we further provide a sampling-based method to approximate the similarity score over the entity pair space for computational efficiency. Besides developing a framework for assessing the similarity between relations, our second contribution is that we have done a survey of applications. We present experiments and analysis aimed at answering five questions: (1) How well does the computed similarity score correlate with human judgment about the similarity between relations? How does our approach compare to other possible approaches based on other kinds of relation embeddings to define a similarity? (sec:relationship and sec:human-judgment) (2) Open IE models inevitably extract many redundant relations. How can our approach help reduce such redundancy? (sec:openie) (3) To which extent, quantitatively, does best relational classification models make errors among similar relations? (sec:error-analysis) (4) Could similarity be used in a heuristic method to enhance negative sampling for relation prediction? (sec:training-guidance-relation-prediction) (5) Could similarity be used as an adaptive margin in softmax-margin training method for relation extraction? (sec:training-guidance-relation-extraction) Finally, we conclude with a discussion of valid extensions to our method and other possible applications. ### Learning Head-Tail Distribution
Just as introduced in sec:introduction, we quantify the similarity between relations by their corresponding head-tail entity pair distributions. Consider the typical case that we have got numbers of facts, but they are still sparse among all facts in the real world. How could we obtain a well-generalized distribution over the whole space of possible triples beyond the training facts? This section proposes a method to parameterize such a distribution. ### Formal Definition of Fact Distribution
A fact is a triple INLINEFORM0 , where INLINEFORM1 and INLINEFORM2 are called head and tail entities, INLINEFORM3 is the relation connecting them, INLINEFORM4 and INLINEFORM5 are the sets of entities and relations respectively. We consider a score function INLINEFORM6 maps all triples to a scalar value. As a special case, the function can be factorized into the sum of two parts: INLINEFORM7 . We use INLINEFORM8 to define the unnormalized probability. DISPLAYFORM0 for every triple INLINEFORM0 . The real parameter INLINEFORM1 can be adjusted to obtain difference distributions over facts. In this paper, we only consider locally normalized version of INLINEFORM0 : DISPLAYFORM0 where INLINEFORM0 and INLINEFORM1 are directly parameterized by feed-forward neural networks. Through local normalization, INLINEFORM2 is naturally a valid probability distribution, as the partition function INLINEFORM3 . Therefore, INLINEFORM4 . ### Neural architecture design
Here we introduce our special design of neural networks. For the first part and the second part, we implement the scoring functions introduced in eq:local-normalization as DISPLAYFORM0 where each INLINEFORM0 represents a multi-layer perceptron composed of layers like INLINEFORM1 , INLINEFORM2 , INLINEFORM3 are embeddings of INLINEFORM4 , INLINEFORM5 , and INLINEFORM6 includes weights and biases in all layers. ### Training
Now we discuss the method to perform training. In this paper, we consider joint training. By minimizing the loss function, we compute the model parameters INLINEFORM0 : DISPLAYFORM0 where INLINEFORM0 is a set of triples. The whole set of parameters, INLINEFORM1 . We train these parameters by Adam optimizer BIBREF2 . Training details are shown in sec:trainingdetail. ### Quantifying Similarity
So far, we have talked about how to use neural networks to approximate the natural distribution of facts. The center topic of our paper, quantifying similarity, will be discussed in detail in this section. ### Relations as Distributions
In this paper, we provide a probability view of relations by representing relation INLINEFORM0 as a probability distribution INLINEFORM1 . After training the neural network on a given set of triples, the model is expected to generalize well on the whole INLINEFORM2 space. Note that it is very easy to calculate INLINEFORM0 in our model thanks to local normalization (eq:local-normalization). Therefore, we can compute it by DISPLAYFORM0 ### Defining Similarity
As the basis of our definition, we hypothesize that the similarity between INLINEFORM0 reflects the similarity between relations. For example, if the conditional distributions of two relations put mass on similar entity pairs, the two relations should be quite similar. If they emphasize different ones, the two should have some differences in meaning. Formally, we define the similarity between two relations as a function of the divergence between the distributions of corresponding head-tail entity pairs: DISPLAYFORM0 where INLINEFORM0 denotes Kullback–Leibler divergence, DISPLAYFORM0 vice versa, and function INLINEFORM0 is a symmetrical function. To keep the coherence between semantic meaning of “similarity” and our definition, INLINEFORM1 should be a monotonically decreasing function. Through this paper, we choose to use an exponential family composed with max function, i.e., INLINEFORM2 . Note that by taking both sides of KL divergence into account, our definition incorporates both the entity pairs with high probability in INLINEFORM3 and INLINEFORM4 . Intuitively, if INLINEFORM5 mainly distributes on a proportion of entities pairs that INLINEFORM6 emphasizes, INLINEFORM7 is only hyponymy of INLINEFORM8 . Considering both sides of KL divergence could help model yield more comprehensive consideration. We will talk about the advantage of this method in detail in sec:relationship. ### Calculating Similarity
Just as introduced in sec:introduction, it is intractable to compute similarity exactly, as involving INLINEFORM0 computation. Hence, we consider the monte-carlo approximation: DISPLAYFORM0 where INLINEFORM0 is a list of entity pairs sampled from INLINEFORM1 . We use sequential sampling to gain INLINEFORM6 , which means we first sample INLINEFORM7 given INLINEFORM8 from INLINEFORM9 , and then sample INLINEFORM10 given INLINEFORM11 and INLINEFORM12 from INLINEFORM13 . ### Relationship with other metrics
Previous work proposed various methods for representing relations as vectors BIBREF3 , BIBREF4 , as matrices BIBREF5 , even as angles BIBREF6 , etc. Based on each of these representations, one could easily define various similarity quantification methods. We show in tab:other-similarity the best one of them in each category of relation presentation. Here we provide two intuitive reasons for using our proposed probability-based similarity: (1) the capacity of a single fixed-size representation is limited — some details about the fact distribution is lost during embedding; (2) directly comparing distributions yields a better interpretability — you can not know about how two relations are different given two relation embeddings, but our model helps you study the detailed differences between probabilities on every entity pair. fig:head-tail-distribution provides an example. Although the two relations talk about the same topic, they have different meanings. TransE embeds them as vectors the closest to each other, while our model can capture the distinction between the distributions corresponds to the two relations, which could be directly noticed from the figure. Embeddings used in this graph are from a trained TransE model. ### Dataset Construction
We show the statistics of the dataset we use in tab:statistics, and the construction procedures will be introduced in this section. ### Wikidata
In Wikidata BIBREF8 , facts can be described as (Head item/property, Property, Tail item/property). To construct a dataset suitable for our task, we only consider the facts whose head entity and tail entity are both items. We first choose the most common 202 relations and 120000 entities from Wikidata as our initial data. Considering that the facts containing the two most frequently appearing relations (P2860: cites, and P31: instance of) occupy half of the initial data, we drop the two relations to downsize the dataset and make the dataset more balanced. Finally, we keep the triples whose head and tail both come from the selected 120000 entities as well as its relation comes from the remaining 200 relations. ### ReVerb Extractions
ReVerb BIBREF9 is a program that automatically identifies and extracts binary relationships from English sentences. We use the extractions from running ReVerb on Wikipedia. We only keep the relations appear more than 10 times and their corresponding triples to construct our dataset. ### FB15K and TACRED
FB15K BIBREF3 is a subset of freebase. TACRED BIBREF10 is a large supervised relation extraction dataset obtained via crowdsourcing. We directly use these two dataset, no extra processing steps were applied. ### Human Judgments
Following BIBREF11 , BIBREF12 and the vast amount of previous work on semantic similarity, we ask nine undergraduate subjects to assess the similarity of 360 pairs of relations from a subset of Wikidata BIBREF8 that are chosen to cover from high to low levels of similarity. In our experiment, subjects were asked to rate an integer similarity score from 0 (no similarity) to 4 (perfectly the same) for each pair. The inter-subject correlation, estimated by leaving-one-out method BIBREF13 , is r = INLINEFORM0 , standard deviation = INLINEFORM1 . This important reference value (marked in fig:correlation) could be seen as the highest expected performance for machines BIBREF12 . To get baselines for comparison, we consider other possible methods to define similarity functions, as shown in tab:other-similarity. We compute the correlation between these methods and human judgment scores. As the models we have chosen are the ones work best in knowledge base completion, we do expect the similarity quantification approaches based on them could measure some degree of similarity. As shown in fig:correlation, the three baseline models could achieve moderate ( INLINEFORM0 ) positive correlation. On the other hand, our model shows a stronger correlation ( INLINEFORM1 ) with human judgment, indicating that considering the probability over whole entity pair space helps to gain a similarity closer to human judgments. These results provide evidence for our claim raised in sec:defining-similarity. ### Redundant Relation Removal
Open IE extracts concise token patterns from plain text to represent various relations between entities, e.g.,, (Mark Twain, was born in, Florida). As Open IE is significant for constructing KBs, many effective extractors have been proposed to extract triples, such as Text-Runner BIBREF14 , ReVerb BIBREF9 , and Standford Open IE BIBREF15 . However, these extractors only yield relation patterns between entities, without aggregating and clustering their results. Accordingly, there are a fair amount of redundant relation patterns after extracting those relation patterns. Furthermore, the redundant patterns lead to some redundant relations in KBs. Recently, some efforts are devoted to Open Relation Extraction (Open RE) BIBREF16 , BIBREF17 , BIBREF18 , BIBREF19 , aiming to cluster relation patterns into several relation types instead of redundant relation patterns. Whenas, these Open RE methods adopt distantly supervised labels as golden relation types, suffering from both false positive and false negative problems on the one hand. On the other hand, these methods still rely on the conventional similarity metrics mentioned above. In this section, we will show that our defined similarity quantification could help Open IE by identifying redundant relations. To be specific, we set a toy experiment to remove redundant relations in KBs for a preliminary comparison (sec:toy-experiment). Then, we evaluate our model and baselines on the real-world dataset extracted by Open IE methods (sec:real-experiment). Considering the existing evaluation metric for Open IE and Open RE rely on either labor-intensive annotations or distantly supervised annotations, we propose a metric approximating recall and precision evaluation based on operable human annotations for balancing both efficiency and accuracy. ### Toy Experiment
In this subsection, we propose a toy environment to verify our similarity-based method. Specifically, we construct a dataset from Wikidata and implement Chinese restaurant process to split every relation in the dataset into several sub-relations. Then, we filter out those sub-relations appearing less than 50 times to eventually get 1165 relations. All these split relations are regarded as different ones during training, and then different relation similarity metrics are adopted to merge those sub-relations into one relation. As Figure FIGREF26 shown that the matrices-based approach is less effective than other approaches, we leave this approach out of this experiment. The results are shown in Table TABREF37 . ### Real World Experiment
In this subsection, we evaluate various relation similarity metrics on the real-world Open IE patterns. The dataset are constructed by ReVerb. Different patterns will be regarded as different relations during training, and we also adopt various relation similarity metrics to merge similar relation patterns. Because it is nearly impossible to annotate all pattern pairs for their merging or not, meanwhile it is also inappropriate to take distantly supervised annotations as golden results. Hence, we propose a novel metric approximating recall and precision evaluation based on minimal human annotations for evaluation in this experiment. Recall is defined as the yielding fraction of true positive instances over the total amount of real positive instances. However, we do not have annotations about which pairs of relations are synonymous. Crowdsourcing is a method to obtain a large number of high-quality annotations. Nevertheless, applying crowdsourcing is not trivial in our settings, because it is intractable to enumerate all synonymous pairs in the large space of relation (pattern) pairs INLINEFORM0 in Open IE. A promising method is to use rejection sampling by uniform sampling from the whole space, and only keep the synonymous ones judged by crowdworkers. However, this is not practical either, as the synonymous pairs are sparse in the whole space, resulting in low efficiency. Fortunately, we could use normalized importance sampling as an alternative to get an unbiased estimation of recall. Theorem 1 Suppose every sample INLINEFORM0 has a label INLINEFORM1 , and the model to be evaluated also gives its prediction INLINEFORM2 . The recall can be written as DISPLAYFORM0 where INLINEFORM0 is the uniform distribution over all samples with INLINEFORM1 . If we have a proposal distribution INLINEFORM2 satisfying INLINEFORM3 , we get an unbiased estimation of recall: DISPLAYFORM0 where INLINEFORM0 is a normalized version of INLINEFORM1 , where INLINEFORM2 is the unnormalized version of q, and INLINEFORM3 are i.i.d. drawn from INLINEFORM4 . Similar to eq:recall-expectation, we can write the expectation form of precision: DISPLAYFORM0 where INLINEFORM0 is the uniform distribution over all samples with INLINEFORM1 . As these samples could be found out by performing models on it. We can simply approximate precision by Monte Carlo Sampling: DISPLAYFORM0 where INLINEFORM0 . In our setting, INLINEFORM0 , INLINEFORM1 means INLINEFORM2 and INLINEFORM3 are the same relations, INLINEFORM4 means INLINEFORM5 is larger than a threshold INLINEFORM6 . The results on the ReVerb Extractions dataset that we constructed are described in fig:precision-recall-openie. To approximate recall, we use the similarity scores as the proposal distribution INLINEFORM0 . 500 relation pairs are then drawn from INLINEFORM1 . To approximate precision, we set thresholds at equal intervals. At each threshold, we uniformly sample 50 to 100 relation pairs whose similarity score given by the model is larger than the threshold. We ask 15 undergraduates to judge whether two relations in a relation pair have the same meaning. A relation pair is viewed valid only if 8 of the annotators annotate it as valid. We use the annotations to approximate recall and precision with eq:recall and eq:precision. Apart from the confidential interval of precision shown in the figure, the largest INLINEFORM2 confidential interval among thresholds for recall is INLINEFORM3 . From the result, we could see that our model performs much better than other models' similarity by a very large margin. ### Error Analysis for Relational Classification
In this section, we consider two kinds of relational classification tasks: (1) relation prediction and (2) relation extraction. Relation prediction aims at predicting the relationship between entities with a given set of triples as training data; while relation extraction aims at extracting the relationship between two entities in a sentence. ### Relation Prediction
We hope to design a simple and clear experiment setup to conduct error analysis for relational prediction. Therefore, we consider a typical method TransE BIBREF3 as the subject as well as FB15K BIBREF3 as the dataset. TransE embeds entities and relations as vectors, and train these embeddings by minimizing DISPLAYFORM0 where INLINEFORM0 is the set of training triples, INLINEFORM1 is the distance function, INLINEFORM2 is a negative sample with one element different from INLINEFORM4 uniformly sampled from INLINEFORM5 , and INLINEFORM6 is the margin. During testing, for each entity pair INLINEFORM0 , TransE rank relations according to INLINEFORM1 . For each INLINEFORM2 in the test set, we call the relations with higher rank scores than INLINEFORM3 distracting relations. We then compare the similarity between the golden relation and distracting relations. Note that some entity pairs could correspond to more than one relations, in which case we just do not see them as distracting relations. ### Relation Extraction
For relation extraction, we consider the supervised relation extraction setting and TACRED dataset BIBREF10 . As for the subject model, we use the best model on TACRED dataset — position-aware neural sequence model. This method first passes the sentence into an LSTM and then calculate an attention sum of the hidden states in the LSTM by taking positional features into account. This simple and effective method achieves the best in TACRED dataset. ### Results
fig:averank shows the distribution of similarity ranks of distracting relations of the above mentioned models' outputs on both relation prediction and relation extraction tasks. From fig:averankrp,fig:averankre, we could observe the most distracting relations are the most similar ones, which corroborate our hypothesis that even the best models on these tasks still make mistakes among the most similar relations. This result also highlights the importance of a heuristic method for guiding models to pay more attention to the boundary between similar relations. We also try to do the negative sampling with relation type constraints, but we see no improvement compared with uniform sampling. The details of negative sampling with relation type constraints are presented in sec:relation-type-constraints. ### Similarity and Negative Sampling
Based on the observation presented in sec:erroranalysisresult, we find out that similar relations are often confusing for relation prediction models. Therefore, corrupted triples with similar relations can be used as high-quality negative samples. For a given valid triple INLINEFORM0 , we corrupt the triple by substituting INLINEFORM1 with INLINEFORM2 with the probability, DISPLAYFORM0 where INLINEFORM0 is the temperature of the exponential function, the bigger the INLINEFORM1 is, the flatter the probability distribution is. When the temperature approaches infinite, the sampling process reduces to uniform sampling. In training, we set the initial temperature to a high level and gradually reduce the temperature. Intuitively, it enables the model to distinguish among those obviously different relations in the early stage and gives more and more confusing negative triples as the training processes to help the model distinguish the similar relations. This can be also viewed as a process of curriculum learning BIBREF21 , the data fed to the model gradually changes from simple negative triples to hard ones. We perform relation prediction task on FB15K with TransE. Following BIBREF3 , we use the "Filtered" setting protocol, i.e., filtering out the corrupted triples that appear in the dataset. Our sampling method is shown to improve the model's performance, especially on Hit@1 (fig:relationprediction). Training details are described in sec:trainingdetail. ### Similarity and Softmax-Margin Loss
Similar to sec:training-guidance-relation-prediction, we find out that relation extraction models often make wrong preditions on similar relations. In this section, we use similarity as an adaptive margin in softmax-margin loss to improve the performance of relation extraction models. As shown in BIBREF22 , Softmax-Margin Loss can be expressed as DISPLAYFORM0 where INLINEFORM0 denotes a structured output space for INLINEFORM1 , and INLINEFORM2 is INLINEFORM3 example in training data. We can easily incorporate similarity into cost function INLINEFORM0 . In this task, we define the cost function as INLINEFORM1 , where INLINEFORM2 is a hyperparameter. Intuitively, we give a larger margin between similar relations, forcing the model to distinguish among them, and thus making the model perform better. We apply our method to Position-aware Attention LSTM (PA-LSTM) BIBREF10 , and tab:relationextraction shows our method improves the performance of PA-LSTM. Training details are described in sec:trainingdetail. ### Related Works
As many early works devoted to psychology and linguistics, especially those works exploring semantic similarity BIBREF11 , BIBREF12 , researchers have empirically found there are various different categorizations of semantic relations among words and contexts. For promoting research on these different semantic relations, bejar1991cognitive explicitly defining these relations and miller1995wordnet further systematically organize rich semantic relations between words via a database. For identifying correlation and distinction between different semantic relations so as to support learning semantic similarity, various methods have attempted to measure relational similarity BIBREF23 , BIBREF24 , BIBREF25 , BIBREF26 , BIBREF27 , BIBREF28 , BIBREF29 . With the ongoing development of information extraction and effective construction of KBs BIBREF0 , BIBREF1 , BIBREF30 , relations are further defined as various types of latent connections between objects more than semantic relations. These general relations play a core role in expressing relational facts in the real world. Hence, there are accordingly various methods proposed for discovering more relations and their facts, including open information extraction BIBREF31 , BIBREF32 , BIBREF33 , BIBREF34 , BIBREF35 , BIBREF36 , BIBREF37 and relation extraction BIBREF38 , BIBREF39 , BIBREF40 , BIBREF41 , BIBREF42 , BIBREF43 , and relation prediction BIBREF3 , BIBREF44 , BIBREF45 , BIBREF46 , BIBREF47 . For both semantic relations and general relations, identifying them is a crucial problem, requiring systems to provide a fine-grained relation similarity metric. However, the existing methods suffer from sparse data, which makes it difficult to achieve an effective and stable similarity metric. Motivated by this, we propose to measure relation similarity by leveraging their fact distribution so that we can identify nuances between similar relations, and merge those distant surface forms of the same relations, benefitting the tasks mentioned above. ### Conclusion and Future Work
In this paper, we introduce an effective method to quantify the relation similarity and provide analysis and a survey of applications. We note that there are a wide range of future directions: (1) human prior knowledge could be incorporated into the similarity quantification; (2) similarity between relations could also be considered in multi-modal settings, e.g., extracting relations from images, videos, or even from audios; (3) by analyzing the distributions corresponding to different relations, one can also find some “meta-relations” between relations, such as hypernymy and hyponymy. ### Acknowledgements
This work is supported by the National Natural Science Foundation of China (NSFC No. 61572273, 61532010), the National Key Research and Development Program of China (No. 2018YFB1004503). Chen and Zhu is supported by Tsinghua University Initiative Scientific Research Program, and Chen is also supported by DCST Student Academic Training Program. Han is also supported by 2018 Tencent Rhino-Bird Elite Training Program. ### Proofs to theorems in the paper
If we have a proposal distribution INLINEFORM0 satisfying INLINEFORM1 , then eq:proofrecallfirstpart can be further written as DISPLAYFORM0 Sometimes, it's hard for us to compute normalized probability INLINEFORM0 . To tackle this problem, consider self-normalized importance sampling as an unbiased estimation BIBREF50 , DISPLAYFORM0 where INLINEFORM0 is the normalized version of INLINEFORM1 . ### Chinese Restaurant Process
Specifically, for a relation INLINEFORM0 with currently INLINEFORM1 sub-relations, we turn it to a new sub-relation with probability DISPLAYFORM0 or to the INLINEFORM0 existing sub-relation with probability DISPLAYFORM0 where INLINEFORM0 is the size of INLINEFORM1 existing sub-relation, INLINEFORM2 is the sum of the number of all sub-relationships of INLINEFORM3 , and INLINEFORM4 is a hyperparameter, in which case we use INLINEFORM5 . ### Training Details
In Wikidata and ReVerb Extractions dataset, we manually split a validation set, assuring every entity and relation appears in validation set also appears in training set. While minimizing loss on the training set, we observe the loss on the validation set and stop training as validation loss stops to decrease. Before training our model on any dataset, we use the entity embeddings and relation embeddings produced by TransE on the dataset as the pretrained embeddings for our model. ### Training Details on Negative Sampling
The sampling is launched with an initial temperature of 8192. The temperature drops to half every 200 epochs and remains stable once it hits 16. Optimization is performed using SGD, with a learning rate of 1e-3. ### Training Details on Softmax-Margin Loss
The sampling is launching with an initial temperature of 64. The temperature drops by 20% per epoch, and remains stable once it hits 16. The alpha we use is 9. Optimization is performed using SGD, with a learning rate of 1. ### Recall Standard Deviation
As is shown in fig:recallstd, the max recall standard deviation for our model is 0.4, and 0.11 for TransE. ### Negative Samplilng with Relation Type Constraints
In FB15K, if two relations have same prefix, we regard them as belonging to a same type, e.g., both /film/film/starring./film/performance/actor and /film/actor/film./film/performance/film have prefix film, they belong to same type. Similar to what is mentioned in sec:training-guidance-relation-prediction, we expect the model first to learn to distinguish among obviously different relations, and gradually learn to distinguish similar relations. Therefore, we conduct negative sampling with relation type constraints in two ways. ### Add Up Two Uniform Distribution
For each triple INLINEFORM0 , we have two uniform distribution INLINEFORM1 and INLINEFORM2 . INLINEFORM3 is the uniform distribution over all the relations except for those appear with INLINEFORM4 in the knowledge base, and INLINEFORM5 is the uniform distribution over the relations of the same type as INLINEFORM6 . When corrupting the triple, we sample INLINEFORM7 from the distribution: DISPLAYFORM0 where INLINEFORM0 is a hyperparameter. We set INLINEFORM1 to 1 at the beginning of training, and every INLINEFORM2 epochs, INLINEFORM3 will be multiplied by decrease rate INLINEFORM4 . We do grid search for INLINEFORM5 and INLINEFORM6 , but no improvement is observed. ### Add Weight
We speculate that the unsatisfactory result produced by adding up two uniform distribution is because that for those types with few relations in it, a small change of INLINEFORM0 will result in a significant change in INLINEFORM1 . Therefore, when sampling a negative INLINEFORM2 , we add weights to relations that are of the same type as INLINEFORM3 instead. Concretely, we substitute INLINEFORM4 with INLINEFORM5 with probability INLINEFORM6 , which can be calculated as: DISPLAYFORM0 where INLINEFORM0 denotes all the relations that are the same type as INLINEFORM1 , INLINEFORM2 is a hyperparameter and INLINEFORM3 is a normalizing constant. We set INLINEFORM4 to 0 at the beginning of training, and every INLINEFORM5 epochs, INLINEFORM6 will increase by INLINEFORM7 . We do grid search for INLINEFORM8 and INLINEFORM9 , still no improvement is observed. ### Wikidata annotation guidance
We show the guidance provided for the annotators here. Table 1: An illustration of the errors made by relation extraction models. The sentence contains obvious patterns indicating the two persons are siblings, but the model predicts it as parents. We introduce an approach to measure the similarity between relations. Our result shows “siblings” is the second most similar one to “parents”. By applying this approach, we could analyze the errors made by models, and help reduce errors. Table 2: Methods to define a similarity function with different types of relation representations Figure 1: Head-tail entity pairs of relation “be an unincorporated community in” (in blue) and “be a small city in” (in red) sampled from our fact distribution model. The coordinates of the points are computed by t-sne (Maaten and Hinton, 2008) on the concatenation of head and tail embeddings8. The two larger blue and red points indicate the embeddings of these two relations. Figure 2: Spearman correlations between human judgment and models’ outputs. The inter-subject correlation is also shown as a horizontal line with its standard deviation as an error band. Our model shows the strongest positive correlation with human judgment, and, in other words, the smallest margin with human inter-subject agreement. Significance: ***/**/* := p < .001/.01/.05. Table 3: Statistics of the triple sets used in this paper. Table 4: The experiment results on the toy dataset show that our metric based on probability distribution significantly outperforms other relation similarity metrics. Figure 3: Precision-recall curve on Open IE task comparing our similarity function with vector-based and angle-based similarity. Error bar represents 95% confidential interval. Bootstraping is used to calculate the confidential interval. Figure 4: Similarity rank distributions of distracting relations on different tasks and datasets. Most of the distracting relations have top similarity rank. Distracting relations are, as defined previously, the relations have a higher rank in the relation classification result than the ground truth. Table 5: Improvement of using similarity in softmaxmargin loss. Figure 6: The recall standard deviation of different models. | relation prediction, relation extraction, Open IE |
Did they use the state-of-the-art model to analyze the attention? | ### Introduction
Deep learning has achieved tremendous success for many NLP tasks. However, unlike traditional methods that provide optimized weights for human understandable features, the behavior of deep learning models is much harder to interpret. Due to the high dimensionality of word embeddings, and the complex, typically recurrent architectures used for textual data, it is often unclear how and why a deep learning model reaches its decisions. There are a few attempts toward explaining/interpreting deep learning-based models, mostly by visualizing the representation of words and/or hidden states, and their importances (via saliency or erasure) on shallow tasks like sentiment analysis and POS tagging BIBREF0 , BIBREF1 , BIBREF2 , BIBREF3 . In contrast, we focus on interpreting the gating and attention signals of the intermediate layers of deep models in the challenging task of Natural Language Inference. A key concept in explaining deep models is saliency, which determines what is critical for the final decision of a deep model. So far, saliency has only been used to illustrate the impact of word embeddings. In this paper, we extend this concept to the intermediate layer of deep models to examine the saliency of attention as well as the LSTM gating signals to understand the behavior of these components and their impact on the final decision. We make two main contributions. First, we introduce new strategies for interpreting the behavior of deep models in their intermediate layers, specifically, by examining the saliency of the attention and the gating signals. Second, we provide an extensive analysis of the state-of-the-art model for the NLI task and show that our methods reveal interesting insights not available from traditional methods of inspecting attention and word saliency. In this paper, our focus was on NLI, which is a fundamental NLP task that requires both understanding and reasoning. Furthermore, the state-of-the-art NLI models employ complex neural architectures involving key mechanisms, such as attention and repeated reading, widely seen in successful models for other NLP tasks. As such, we expect our methods to be potentially useful for other natural understanding tasks as well. ### Task and Model
In NLI BIBREF4 , we are given two sentences, a premise and a hypothesis, the goal is to decide the logical relationship (Entailment, Neutral, or Contradiction) between them. Many of the top performing NLI models BIBREF5 , BIBREF6 , BIBREF7 , BIBREF8 , BIBREF9 , BIBREF10 , BIBREF11 , are variants of the ESIM model BIBREF11 , which we choose to analyze in this paper. ESIM reads the sentences independently using LSTM at first, and then applies attention to align/contrast the sentences. Another round of LSTM reading then produces the final representations, which are compared to make the prediction. Detailed description of ESIM can be found in the Appendix. Using the SNLI BIBREF4 data, we train two variants of ESIM, with dimensionality 50 and 300 respectively, referred to as ESIM-50 and ESIM-300 in the remainder of the paper. ### Visualization of Attention and Gating
In this work, we are primarily interested in the internal workings of the NLI model. In particular, we focus on the attention and the gating signals of LSTM readers, and how they contribute to the decisions of the model. ### Attention
Attention has been widely used in many NLP tasks BIBREF12 , BIBREF13 , BIBREF14 and is probably one of the most critical parts that affects the inference decisions. Several pieces of prior work in NLI have attempted to visualize the attention layer to provide some understanding of their models BIBREF5 , BIBREF15 . Such visualizations generate a heatmap representing the similarity between the hidden states of the premise and the hypothesis (Eq. 19 of Appendix). Unfortunately the similarities are often the same regardless of the decision. Let us consider the following example, where the same premise “A kid is playing in the garden”, is paired with three different hypotheses: A kid is taking a nap in the garden A kid is having fun in the garden with her family A kid is having fun in the garden Note that the ground truth relationships are Contradiction, Neutral, and Entailment, respectively. The first row of Fig. 1 shows the visualization of normalized attention for the three cases produced by ESIM-50, which makes correct predictions for all of them. As we can see from the figure, the three attention maps are fairly similar despite the completely different decisions. The key issue is that the attention visualization only allows us to see how the model aligns the premise with the hypothesis, but does not show how such alignment impacts the decision. This prompts us to consider the saliency of attention. The concept of saliency was first introduced in vision for visualizing the spatial support on an image for a particular object class BIBREF16 . In NLP, saliency has been used to study the importance of words toward a final decision BIBREF0 . We propose to examine the saliency of attention. Specifically, given a premise-hypothesis pair and the model's decision $y$ , we consider the similarity between a pair of premise and hypothesis hidden states $e_{ij}$ as a variable. The score of the decision $S(y)$ is thus a function of $e_{ij}$ for all $i$ and $j$ . The saliency of $e_{ij}$ is then defined to be $|\frac{\partial S(y)}{\partial {e_{ij}}}|$ . The second row of Fig. 1 presents the attention saliency map for the three examples acquired by the same ESIM-50 model. Interestingly, the saliencies are clearly different across the examples, each highlighting different parts of the alignment. Specifically, for h1, we see the alignment between “is playing” and “taking a nap” and the alignment of “in a garden” to have the most prominent saliency toward the decision of Contradiction. For h2, the alignment of “kid” and “her family” seems to be the most salient for the decision of Neutral. Finally, for h3, the alignment between “is having fun” and “kid is playing” have the strongest impact toward the decision of Entailment. From this example, we can see that by inspecting the attention saliency, we effectively pinpoint which part of the alignments contribute most critically to the final prediction whereas simply visualizing the attention itself reveals little information. In the previous examples, we study the behavior of the same model on different inputs. Now we use the attention saliency to compare the two different ESIM models: ESIM-50 and ESIM-300. Consider two examples with a shared hypothesis of “A man ordered a book” and premise: John ordered a book from amazon Mary ordered a book from amazon Here ESIM-50 fails to capture the gender connections of the two different names and predicts Neutral for both inputs, whereas ESIM-300 correctly predicts Entailment for the first case and Contradiction for the second. In the first two columns of Fig. 2 (column a and b) we visualize the attention of the two examples for ESIM-50 (left) and ESIM-300 (right) respectively. Although the two models make different predictions, their attention maps appear qualitatively similar. In contrast, columns 3-4 of Fig. 2 (column c and d) present the attention saliency for the two examples by ESIM-50 and ESIM-300 respectively. We see that for both examples, ESIM-50 primarily focused on the alignment of “ordered”, whereas ESIM-300 focused more on the alignment of “John” and “Mary” with “man”. It is interesting to note that ESIM-300 does not appear to learn significantly different similarity values compared to ESIM-50 for the two critical pairs of words (“John”, “man”) and (“Mary”, “man”) based on the attention map. The saliency map, however, reveals that the two models use these values quite differently, with only ESIM-300 correctly focusing on them. ### LSTM Gating Signals
LSTM gating signals determine the flow of information. In other words, they indicate how LSTM reads the word sequences and how the information from different parts is captured and combined. LSTM gating signals are rarely analyzed, possibly due to their high dimensionality and complexity. In this work, we consider both the gating signals and their saliency, which is computed as the partial derivative of the score of the final decision with respect to each gating signal. Instead of considering individual dimensions of the gating signals, we aggregate them to consider their norm, both for the signal and for its saliency. Note that ESIM models have two LSTM layers, the first (input) LSTM performs the input encoding and the second (inference) LSTM generates the representation for inference. In Fig. 3 we plot the normalized signal and saliency norms for different gates (input, forget, output) of the Forward input (bottom three rows) and inference (top three rows) LSTMs. These results are produced by the ESIM-50 model for the three examples of Section 3.1, one for each column. From the figure, we first note that the saliency tends to be somewhat consistent across different gates within the same LSTM, suggesting that we can interpret them jointly to identify parts of the sentence important for the model's prediction. Comparing across examples, we see that the saliency curves show pronounced differences across the examples. For instance, the saliency pattern of the Neutral example is significantly different from the other two examples, and heavily concentrated toward the end of the sentence (“with her family”). Note that without this part of the sentence, the relationship would have been Entailment. The focus (evidenced by its strong saliency and strong gating signal) on this particular part, which presents information not available from the premise, explains the model's decision of Neutral. Comparing the behavior of the input LSTM and the inference LSTM, we observe interesting shifts of focus. In particular, we see that the inference LSTM tends to see much more concentrated saliency over key parts of the sentence, whereas the input LSTM sees more spread of saliency. For example, for the Contradiction example, the input LSTM sees high saliency for both “taking” and “in”, whereas the inference LSTM primarily focuses on “nap”, which is the key word suggesting a Contradiction. Note that ESIM uses attention between the input and inference LSTM layers to align/contrast the sentences, hence it makes sense that the inference LSTM is more focused on the critical differences between the sentences. This is also observed for the Neutral example as well. It is worth noting that, while revealing similar general trends, the backward LSTM can sometimes focus on different parts of the sentence (e.g., see Fig. 11 of Appendix), suggesting the forward and backward readings provide complementary understanding of the sentence. ### Conclusion
We propose new visualization and interpretation strategies for neural models to understand how and why they work. We demonstrate the effectiveness of the proposed strategies on a complex task (NLI). Our strategies are able to provide interesting insights not achievable by previous explanation techniques. Our future work will extend our study to consider other NLP tasks and models with the goal of producing useful insights for further improving these models. Model In this section we describe the ESIM model. We divide ESIM to three main parts: 1) input encoding, 2) attention, and 3) inference. Figure 4 demonstrates a high-level view of the ESIM framework. Let $u=[u_1, \cdots , u_n]$ and $v=[v_1, \cdots , v_m]$ be the given premise with length $n$ and hypothesis with length $m$ respectively, where $u_i, v_j \in \mathbb {R}^r$ are word embeddings of $r$ -dimensional vector. The goal is to predict a label $y$ that indicates the logical relationship between premise $u$ and hypothesis $v$ . Below we briefly explain the aforementioned parts. Input Encoding It utilizes a bidirectional LSTM (BiLSTM) for encoding the given premise and hypothesis using Equations 16 and 17 respectively. $$\hat{u} \in \mathbb {R}^{n \times 2d}$$ (Eq. ) $$\hat{v} \in \mathbb {R}^{m \times 2d}$$ (Eq. ) where $u$ and $v=[v_1, \cdots , v_m]$0 are the reading sequences of $v=[v_1, \cdots , v_m]$1 and $v=[v_1, \cdots , v_m]$2 respectively. Attention It employs a soft alignment method to associate the relevant sub-components between the given premise and hypothesis. Equation 19 (energy function) computes the unnormalized attention weights as the similarity of hidden states of the premise and hypothesis. $$u$$ (Eq. ) where $v=[v_1, \cdots , v_m]$3 and $v=[v_1, \cdots , v_m]$4 are the hidden representations of $v=[v_1, \cdots , v_m]$5 and $v=[v_1, \cdots , v_m]$6 respectively which are computed earlier in Equations 16 and 17 . Next, for each word in either premise or hypothesis, the relevant semantics in the other sentence is extracted and composed according to $v=[v_1, \cdots , v_m]$7 . Equations 20 and 21 provide formal and specific details of this procedure. $$\tilde{v}_j$$ (Eq. ) $$\hat{u}$$ (Eq. ) where $v=[v_1, \cdots , v_m]$8 represents the extracted relevant information of $v=[v_1, \cdots , v_m]$9 by attending to $n$0 while $n$1 represents the extracted relevant information of $n$2 by attending to $n$3 . Next, it passes the enriched information through a projector layer which produce the final output of attention stage. Equations 22 and 23 formally represent this process. $$p$$ (Eq. ) $$q$$ (Eq. ) Here $n$4 stands for element-wise product while $n$5 and $n$6 are the trainable weights and biases of the projector layer respectively. $n$7 and $n$8 indicate the output of attention devision for premise and hypothesis respectively. Inference During this phase, it uses another BiLSTM to aggregate the two sequences of computed matching vectors, $n$9 and $m$0 from the attention stage (Equations 27 and 28 ). $$\emph {softmax}$$ (Eq. ) $$\hat{u} = \textit {BiLSTM}(u)$$ (Eq. 16) where $m$1 and $m$2 are the reading sequences of $m$3 and $m$4 respectively. Finally the concatenation max and average pooling of $m$5 and $m$6 are pass through a multilayer perceptron (MLP) classifier that includes a hidden layer with $m$7 activation and $m$8 output layer. The model is trained in an end-to-end manner. Attention Study Here we provide more examples on the NLI task which intend to examine specific behavior in this model. Such examples indicate interesting observation that we can analyze them in the future works. Table 1 shows the list of all example. LSTM Gating Signal Finally, Figure 11 depicts the backward LSTM gating signals study. Figure 1: Normalized attention and attention saliency visualization. Each column shows visualization of one sample. Top plots depict attention visualization and bottom ones represent attention saliency visualization. Predicted (the same as Gold) label of each sample is shown on top of each column. | we provide an extensive analysis of the state-of-the-art model |
We can assume that Saladin's army represents which group?
A. Mercenaries
B. Muslims
C. Africans
D. Christians
| ... After a Few Words ... by Seaton McKettrig Illustrated by Summer [Transcriber's Note: This etext was produced from Analog October 1962. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed.] This is a science-fiction story. History is a science; the other part is, as all Americans know, the most fictional field we have today. He settled himself comfortably in his seat, and carefully put the helmet on, pulling it down firmly until it was properly seated. For a moment, he could see nothing. Then his hand moved up and, with a flick of the wrist, lifted the visor. Ahead of him, in serried array, with lances erect and pennons flying, was the forward part of the column. Far ahead, he knew, were the Knights Templars, who had taken the advance. Behind the Templars rode the mailed knights of Brittany and Anjou. These were followed by King Guy of Jerusalem and the host of Poitou. He himself, Sir Robert de Bouain, was riding with the Norman and English troops, just behind the men of Poitou. Sir Robert turned slightly in his saddle. To his right, he could see the brilliant red-and-gold banner of the lion-hearted Richard of England— gules, in pale three lions passant guardant or . Behind the standard-bearer, his great war horse moving with a steady, measured pace, his coronet of gold on his steel helm gleaming in the glaring desert sun, the lions of England on his firm-held shield, was the King himself. Further behind, the Knights Hospitallers protected the rear, guarding the column of the hosts of Christendom from harassment by the Bedouins. "By our Lady!" came a voice from his left. "Three days out from Acre, and the accursed Saracens still elude us." Sir Robert de Bouain twisted again in his saddle to look at the knight riding alongside him. Sir Gaeton de l'Arc-Tombé sat tall and straight in his saddle, his visor up, his blue eyes narrowed against the glare of the sun. Sir Robert's lips formed a smile. "They are not far off, Sir Gaeton. They have been following us. As we march parallel to the seacoast, so they have been marching with us in those hills to the east." "Like the jackals they are," said Sir Gaeton. "They assail us from the rear, and they set up traps in our path ahead. Our spies tell us that the Turks lie ahead of us in countless numbers. And yet, they fear to face us in open battle." "Is it fear, or are they merely gathering their forces?" "Both," said Sir Gaeton flatly. "They fear us, else they would not dally to amass so fearsome a force. If, as our informers tell us, there are uncounted Turks to the fore, and if, as we are aware, our rear is being dogged by the Bedouin and the black horsemen of Egypt, it would seem that Saladin has at hand more than enough to overcome us, were they all truly Christian knights." "Give them time. We must wait for their attack, sir knight. It were foolhardy to attempt to seek them in their own hills, and yet they must stop us. They will attack before we reach Jerusalem, fear not." "We of Gascony fear no heathen Musselman," Sir Gaeton growled. "It's this Hellish heat that is driving me mad." He pointed toward the eastern hills. "The sun is yet low, and already the heat is unbearable." Sir Robert heard his own laugh echo hollowly within his helmet. "Perhaps 'twere better to be mad when the assault comes. Madmen fight better than men of cooler blood." He knew that the others were baking inside their heavy armor, although he himself was not too uncomfortable. Sir Gaeton looked at him with a smile that held both irony and respect. "In truth, sir knight, it is apparent that you fear neither men nor heat. Nor is your own blood too cool. True, I ride with your Normans and your English and your King Richard of the Lion's Heart, but I am a Gascon, and have sworn no fealty to him. But to side with the Duke of Burgundy against King Richard—" He gave a short, barking laugh. "I fear no man," he went on, "but if I had to fear one, it would be Richard of England." Sir Robert's voice came like a sword: steely, flat, cold, and sharp. "My lord the King spoke in haste. He has reason to be bitter against Philip of France, as do we all. Philip has deserted the field. He has returned to France in haste, leaving the rest of us to fight the Saracen for the Holy Land leaving only the contingent of his vassal the Duke of Burgundy to remain with us." "Richard of England has never been on the best of terms with Philip Augustus," said Sir Gaeton. "No, and with good cause. But he allowed his anger against Philip to color his judgment when he spoke harshly against the Duke of Burgundy. The Duke is no coward, and Richard Plantagenet well knows it. As I said, he spoke in haste." "And you intervened," said Sir Gaeton. "It was my duty." Sir Robert's voice was stubborn. "Could we have permitted a quarrel to develop between the two finest knights and warleaders in Christendom at this crucial point? The desertion of Philip of France has cost us dearly. Could we permit the desertion of Burgundy, too?" "You did what must be done in honor," the Gascon conceded, "but you have not gained the love of Richard by doing so." Sir Robert felt his jaw set firmly. "My king knows I am loyal." Sir Gaeton said nothing more, but there was a look in his eyes that showed that he felt that Richard of England might even doubt the loyalty of Sir Robert de Bouain. Sir Robert rode on in silence, feeling the movement of the horse beneath him. There was a sudden sound to the rear. Like a wash of the tide from the sea came the sound of Saracen war cries and the clash of steel on steel mingled with the sounds of horses in agony and anger. Sir Robert turned his horse to look. The Negro troops of Saladin's Egyptian contingent were thundering down upon the rear! They clashed with the Hospitallers, slamming in like a rain of heavy stones, too close in for the use of bows. There was only the sword against armor, like the sound of a thousand hammers against a thousand anvils. "Stand fast! Stand fast! Hold them off!" It was the voice of King Richard, sounding like a clarion over the din of battle. Sir Robert felt his horse move, as though it were urging him on toward the battle, but his hand held to the reins, keeping the great charger in check. The King had said "Stand fast!" and this was no time to disobey the orders of Richard. The Saracen troops were coming in from the rear, and the Hospitallers were taking the brunt of the charge. They fought like madmen, but they were slowly being forced back. The Master of the Hospitallers rode to the rear, to the King's standard, which hardly moved in the still desert air, now that the column had stopped moving. The voice of the Duke of Burgundy came to Sir Robert's ears. "Stand fast. The King bids you all to stand fast," said the duke, his voice fading as he rode on up the column toward the knights of Poitou and the Knights Templars. The Master of the Hospitallers was speaking in a low, urgent voice to the King: "My lord, we are pressed on by the enemy and in danger of eternal infamy. We are losing our horses, one after the other!" "Good Master," said Richard, "it is you who must sustain their attack. No one can be everywhere at once." The Master of the Hospitallers nodded curtly and charged back into the fray. The King turned to Sir Baldwin de Carreo, who sat ahorse nearby, and pointed toward the eastern hills. "They will come from there, hitting us in the flank; we cannot afford to amass a rearward charge. To do so would be to fall directly into the hands of the Saracen." A voice very close to Sir Robert said: "Richard is right. If we go to the aid of the Hospitallers, we will expose the column to a flank attack." It was Sir Gaeton. "My lord the King," Sir Robert heard his voice say, "is right in all but one thing. If we allow the Egyptians to take us from the rear, there will be no need for Saladin and his Turks to come down on our flank. And the Hospitallers cannot hold for long at this rate. A charge at full gallop would break the Egyptian line and give the Hospitallers breathing time. Are you with me?" "Against the orders of the King?" "The King cannot see everything! There are times when a man must use his own judgment! You said you were afraid of no man. Are you with me?" After a moment's hesitation, Sir Gaeton couched his lance. "I'm with you, sir knight! Live or die, I follow! Strike and strike hard!" "Forward then!" Sir Robert heard himself shouting. "Forward for St. George and for England!" "St. George and England!" the Gascon echoed. Two great war horses began to move ponderously forward toward the battle lines, gaining momentum as they went. Moving in unison, the two knights, their horses now at a fast trot, lowered their lances, picking their Saracen targets with care. Larger and larger loomed the Egyptian cavalrymen as the horses changed pace to a thundering gallop. The Egyptians tried to dodge, as they saw, too late, the approach of the Christian knights. Sir Robert felt the shock against himself and his horse as the steel tip of the long ash lance struck the Saracen horseman in the chest. Out of the corner of his eye, he saw that Sir Gaeton, too, had scored. The Saracen, impaled on Sir Robert's lance, shot from the saddle as he died. His lighter armor had hardly impeded the incoming spear-point, and now his body dragged it down as he dropped toward the desert sand. Another Moslem cavalryman was charging in now, swinging his curved saber, taking advantage of Sir Robert's sagging lance. There was nothing else to do but drop the lance and draw his heavy broadsword. His hand grasped it, and it came singing from its scabbard. The Egyptian's curved sword clanged against Sir Robert's helm, setting his head ringing. In return, the knight's broadsword came about in a sweeping arc, and the Egyptian's horse rode on with the rider's headless body. Behind him, Sir Robert heard further cries of "St. George and England!" The Hospitallers, taking heart at the charge, were going in! Behind them came the Count of Champagne, the Earl of Leister, and the Bishop of Beauvais, who carried a great warhammer in order that he might not break Church Law by shedding blood. Sir Robert's own sword rose and fell, cutting and hacking at the enemy. He himself felt a dreamlike detachment, as though he were watching the battle rather than participating in it. But he could see that the Moslems were falling back before the Christian onslaught. And then, quite suddenly, there seemed to be no foeman to swing at. Breathing heavily, Sir Robert sheathed his broadsword. Beside him, Sir Gaeton did the same, saying: "It will be a few minutes before they can regroup, sir knight. We may have routed them completely." "Aye. But King Richard will not approve of my breaking ranks and disobeying orders. I may win the battle and lose my head in the end." "This is no time to worry about the future," said the Gascon. "Rest for a moment and relax, that you may be the stronger later. Here—have an Old Kings ." He had a pack of cigarettes in his gauntleted hand, which he profferred to Sir Robert. There were three cigarettes protruding from it, one slightly farther than the others. Sir Robert's hand reached out and took that one. "Thanks. When the going gets rough, I really enjoy an Old Kings ." He put one end of the cigarette in his mouth and lit the other from the lighter in Sir Gaeton's hand. "Yes, sir," said Sir Gaeton, after lighting his own cigarette, " Old Kings are the greatest. They give a man real, deep-down smoking pleasure." "There's no doubt about it, Old Kings are a man's cigarette." Sir Robert could feel the soothing smoke in his lungs as he inhaled deeply. "That's great. When I want a cigarette, I don't want just any cigarette." "Nor I," agreed the Gascon. " Old Kings is the only real cigarette when you're doing a real man's work." "That's for sure." Sir Robert watched a smoke ring expand in the air. There was a sudden clash of arms off to their left. Sir Robert dropped his cigarette to the ground. "The trouble is that doing a real he-man's work doesn't always allow you to enjoy the fine, rich tobaccos of Old Kings right down to the very end." "No, but you can always light another later," said the Gascon knight. King Richard, on seeing his army moving suddenly toward the harassed rear, had realized the danger and had charged through the Hospitallers to get into the thick of the fray. Now the Turks were charging down from the hills, hitting—not the flank as he had expected, but the rear! Saladin had expected him to hold fast! Sir Robert and Sir Gaeton spurred their chargers toward the flapping banner of England. The fierce warrior-king of England, his mighty sword in hand, was cutting down Turks as though they were grain-stalks, but still the Saracen horde pressed on. More and more of the terrible Turks came boiling down out of the hills, their glittering scimitars swinging. Sir Robert lost all track of time. There was nothing to do but keep his own great broadsword moving, swinging like some gigantic metronome as he hacked down the Moslem foes. And then, suddenly, he found himself surrounded by the Saracens! He was isolated and alone, cut off from the rest of the Christian forces! He glanced quickly around as he slashed another Saracen from pate to breastbone. Where was Sir Gaeton? Where were the others? Where was the red-and-gold banner of Richard? He caught a glimpse of the fluttering banner far to the rear and started to fall back. And then he saw another knight nearby, a huge man who swung his sparkling blade with power and force. On his steel helm gleamed a golden coronet! Richard! And the great king, in spite of his prowess was outnumbered heavily and would, within seconds, be cut down by the Saracen horde! Without hesitation, Sir Robert plunged his horse toward the surrounded monarch, his great blade cutting a path before him. He saw Richard go down, falling from the saddle of his charger, but by that time his own sword was cutting into the screaming Saracens and they had no time to attempt any further mischief to the King. They had their hands full with Sir Robert de Bouain. He did not know how long he fought there, holding his charger motionless over the inert body of the fallen king, hewing down the screaming enemy, but presently he heard the familiar cry of "For St. George and for England" behind him. The Norman and English troops were charging in, bringing with them the banner of England! And then Richard was on his feet, cleaving the air about him with his own broadsword. Its bright edge, besmeared with Saracen blood, was biting viciously into the foe. The Turks began to fall back. Within seconds, the Christian knights were boiling around the embattled pair, forcing the Turks into retreat. And for the second time, Sir Robert found himself with no one to fight. And then a voice was saying: "You have done well this day, sir knight. Richard Plantagenet will not forget." Sir Robert turned in his saddle to face the smiling king. "My lord king, be assured that I would never forget my loyalty to my sovereign and liege lord. My sword and my life are yours whenever you call." King Richard's gauntleted hand grasped his own. "If it please God, I shall never ask your life. An earldom awaits you when we return to England, sir knight." And then the king mounted his horse and was running full gallop after the retreating Saracens. Robert took off his helmet. He blinked for a second to adjust his eyes to the relative dimness of the studio. After the brightness of the desert that the televicarion helmet had projected into his eyes, the studio seemed strangely cavelike. "How'd you like it, Bob?" asked one of the two producers of the show. Robert Bowen nodded briskly and patted the televike helmet. "It was O.K.," he said. "Good show. A little talky at the beginning, and it needs a better fade-out, but the action scenes were fine. The sponsor ought to like it—for a while, at least." "What do you mean, 'for a while'?" Robert Bowen sighed. "If this thing goes on the air the way it is, he'll lose sales." "Why? Commercial not good enough?" " Too good! Man, I've smoked Old Kings , and, believe me, the real thing never tasted as good as that cigarette did in the commercial!" | B. Muslims |
Why did the author discuss the movies in this text?
A. they're all based on real-world events
B. they're all meant to improve our views on historical events
C. they all had famous, excellent actors
D. they're all well-written by famous screenwriters
| War and Pieces No movie in the last decade has succeeded in psyching out critics and audiences as fully as the powerful, rambling war epic The Thin Red Line , Terrence Malick's return to cinema after 20 years. I've sat through it twice and am still trying to sort out my responses, which run from awe to mockery and back. Like Saving Private Ryan , the picture wallops you in the gut with brilliant, splattery battle montages and Goyaesque images of hell on earth. But Malick, a certified intellectual and the Pynchonesque figure who directed Badlands and Days of Heaven in the 1970s and then disappeared, is in a different philosophical universe from Steven Spielberg. Post-carnage, his sundry characters philosophize about their experiences in drowsy, runic voice-overs that come at you like slow bean balls: "Why does nature vie with itself? ... Is there an avenging power in nature, not one power but two?" Or "This great evil: Where's it come from? What seed, what root did it grow from? Who's doin' this? Who's killin' us, robbin' us of life and light?" First you get walloped with viscera, then you get beaned by blather. Those existential speculations don't derive from the screenplay's source, an archetypal but otherwise down-to-earth 1962 novel by James Jones (who also wrote From Here to Eternity ) about the American invasion of the South Pacific island of Guadalcanal. They're central to Malick's vision of the story, however, and not specious. In the combat genre, the phrase "war is hell" usually means nothing more than that it's a bummer to lose a limb or two, or to see your buddy get his head blown off. A true work of art owes us more than literal horrors, and Malick obliges by making his theater of war the setting for nothing less than a meditation on the existence of God. He tells the story solemnly, in three parts, with a big-deal cast (Sean Penn, Nick Nolte, John Cusack) and a few other major stars (John Travolta, Woody Harrelson, George Clooney) dropping by for cameos. After an Edenic prelude, in which a boyishly idealistic absent without leave soldier, Pvt. Witt (Jim Caviezel), swims with native youths to the accompaniment of a heavenly children's choir, the first part sees the arrival of the Allied forces on the island, introduces the principal characters (none of whom amounts to a genuine protagonist), and lays out the movie's geographical and philosophical terrain. The centerpiece--the fighting--goes on for over an hour and features the most frantic and harrowing sequences, chiefly the company's initially unsuccessful frontal assault on a Japanese hilltop bunker. The coda lasts nearly 40 minutes and is mostly talk and cleanup, the rhythms growing more relaxed until a final, incongruous spasm of violence--whereupon the surviving soldiers pack their gear and motor off to another South Pacific battle. In the final shot, a twisted tree grows on the waterline of the beach, the cycle of life beginning anew. The Thin Red Line has a curious sound-scape, as the noise of battle frequently recedes to make room for interior monologues and Hans Zimmer's bump-bump, minimalist New Age music. Pvt. Bell (Ben Chaplin) talks to his curvy, redheaded wife, viewed in deliriously sensual flashbacks. ("Love: Where does it come from? Who lit this flame in us?") Lt. Col. Tall (Nolte), a borderline lunatic passed over one too many times for promotion and itching to win a battle no matter what the human cost, worries groggily about how his men perceive him. The dreamer Witt poses folksy questions about whether we're all a part of one big soul. If the movie has a spine, it's his off-and-on dialogue with Sgt. Welsh (Penn), who's increasingly irritated by the private's beatific, almost Billy Budd-like optimism. Says Welsh, "In this world, a man himself is nothin', and there ain't no world but this one." Replies Witt, high cheekbones glinting, "I seen another world." At first it seems as if Witt will indeed be Billy Budd to Welsh's vindictive Claggart. But if Witt is ultimately an ethereal martyr, Welsh turns out to be a Bogart-like romantic who can't stop feeling pain in the face of an absent God. He speaks the movie's epitaph, "Darkness and light, strife and love: Are they the workings of one mind, the feature of the same face? O my soul, let me be in you now. Look out through my eyes. Look out at the things you made, all things shining." Malick puts a lot of shining things on the screen: soldiers, natives, parrots, bats, rodents, visions of Eden by way of National Geographic and of the Fall by way of Alpo. Malick's conception of consciousness distributes it among the animate and inanimate alike; almost every object is held up for rapturous contemplation. I could cite hundreds of images: A soldier in a rocking boat hovers over a letter he's writing, which is crammed from top to bottom and side to side with script. (You don't know the man, but you can feel in an instant his need to cram everything in.) A small, white-bearded Melanesian man strolls nonchalantly past a platoon of tensely trudging grunts who can't believe they're encountering this instead of a hail of Japanese bullets. Two shots bring down the first pair of soldiers to advance on the hill; a second later, the sun plays mystically over the tall, yellow grass that has swallowed their bodies. John Toll's camera rushes in on a captured Japanese garrison: One Japanese soldier shrieks; another, skeletal, laughs and laughs; a third weeps over a dying comrade. The face of a Japanese soldier encased in earth speaks from the dead, "Are you righteous? Know that I was, too." Whether or not these pearllike epiphanies are strung is another matter. Malick throws out his overarching theme--is nature two-sided, at war with itself?--in the first few minutes but, for all his startling juxtapositions, he never dramatizes it with anything approaching the clarity of, say, Brian De Palma's Casualties of War (1989). Besides the dialogue between Welsh and Witt, The Thin Red Line 's other organizing story involves a wrenching tug of war between Nolte's ambition-crazed Tall and Capt. Staros (Elias Koteas), who refuses an order to send his men on what will surely be a suicidal--and futile--assault on a bunker. But matters of cause and effect don't really interest Malick. Individual acts of conscience can and do save lives, and heroism can win a war or a battle, he acknowledges. But Staros is ultimately sent packing, and Malick never bothers to trace the effect of his action on the Guadalcanal operation. In fact, the entire battle seems to take place in a crazed void. Tall quotes Homer's "rosy-fingered dawn" and orders a meaningless bombardment to "buck the men up--it'll look like the Japs are catching hell." Soldiers shoot at hazy figures, unsure whether they're Japanese or American. Men collide, blow themselves in half with their own mishandled grenades, stab themselves frantically with morphine needles, shove cigarettes up their noses to keep the stench of the dying and the dead at bay. A tiny bird, mortally wounded, flutters in the grass. Malick is convincing--at times overwhelming--on the subject of chaos. It's when he tries to ruminate on order that he gets gummed up, retreating to one of his gaseous multiple mouthpieces: "Where is it that we were together? Who is it that I lived with? Walked with? The brother. ... The friend. ... One mind." I think I'd have an easier time with Malick's metaphysical speculations if I had a sense of some concomitant geopolitical ones--central to any larger musings on forces of nature as viewed through the prism of war. Couldn't it be that the German and Japanese fascist orders were profoundly anti-natural, and that the Allies' cause was part of a violent but natural correction? You don't have to buy into Spielberg's Lincolnesque pieties in Saving Private Ryan to believe that there's a difference between World War II and Vietnam (or, for that matter, World War II and the invasion of Grenada or our spats with Iraq). While he was at Harvard, Malick might have peeled himself off the lap of his pointy-headed mentor, Stanley Cavell, the philosopher and film theorist, and checked out a few of Michael Waltzer's lectures on just and unjust wars. Maybe then he'd view Guadalcanal not in an absurdist vacuum (the soldiers come, they kill and are killed, they leave) but in the larger context of a war that was among the most rational (in its aims, if not its methods) fought in the last several centuries. For all his visionary filmmaking, Malick's Zen neutrality sometimes seems like a cultivated--and pretentious--brand of fatuousness. John Travolta's empty nightclub impersonation of Bill Clinton in Primary Colors (1998) had one positive result: It gave him a jump-start on Jan Schlichtmann, the reckless personal injury lawyer at the center of A Civil Action . Travolta's Schlichtmann is much more redolent of Clinton: slick and selfish and corrupt in lots of ways but basically on the side of the angels, too proud and arrogant to change tactics when all is certainly lost. Schlichtmann pursued--and more or less blew--a civil liability case against the corporate giants Beatrice and W.R. Grace over the allegedly carcinogenic water supply of Woburn, Mass. Boston writer Jonathan Harr, in the book the movie is based on, went beyond the poison in the Woburn wells to evoke (stopping just short of libel) the poison of the civil courts, where platoons of overpaid corporate lawyers can drive opponents with pockets less deep and psyches less stable into bankruptcy and hysteria. Director Steven Zaillian's version doesn't capture the mounting rage that one experiences while reading Harr's book, or even the juicy legal machinations that Francis Ford Coppola giddily manipulated in his underrated adaptation of John Grisham's The Rainmaker (1997). But A Civil Action is a sturdy piece of work, an old-fashioned conversion narrative with some high-tech zip. Schlichtmann doesn't take this "orphan" case--brought by the parents of several children who died of leukemia--because he wants to do good but because he figures that Grace and Beatrice will fork over huge sums of money to keep the parents from testifying publicly about their children's last days. He might succeed, too, if it weren't for Jerome Facher (Robert Duvall), the Beatrice lawyer who knows how to keep Schlichtmann shadowboxing while his small firm's financial resources dwindle to nothing. Zaillian is at his most assured when he cuts back and forth between Facher's Harvard Law School lectures on what not to do in court and Schlichtmann's fumbling prosecution. The sequence has the extra dimension of good journalism: It dramatizes and comments simultaneously. Plus, it gives Duvall a splendid platform for impish understatement. (Duvall has become more fun to watch than just about anyone in movies.) Elsewhere, Zaillian takes a more surface approach, sticking to legal minutiae and rarely digging for the deeper evil. As in his Searching for Bobby Fischer (1993), the outcome of every scene is predictable, but how Zaillian gets from beat to beat is surprisingly fresh. He also gets sterling bit performances from Sydney Pollack as the spookily sanguine Grace CEO, William H. Macy as Schlichtmann's rabbity accountant, and Kathleen Quinlan as the mother of one of the victims. Quinlan knows that when you're playing a woman who has lost a child you don't need to emote--you reveal the emotion by trying not to emote. To the families involved in the Woburn tragedy, the real climax of this story isn't the downbeat ending of the book or the sleight of hand, "let's call the Environmental Protection Agency," upbeat ending of the movie. The climax is the publication of a book that takes the plaintiffs' side and that remains on the best-seller list in hardcover and paperback for years. The climax is the movie starring John Travolta. Beatrice and Grace made out OK legally, but some of us will never use their products again without thinking about Travolta losing his shirt in the name of those wasted-away little kids. | A. they're all based on real-world events |
How does Rupert feel about Paul?
A. Rupert thinks of Paul as a kindred spirit.
B. Rupert is annoyed that Paul sat down at his table.
C. Rupert suspects Paul might be a Russian spy.
D. Paul is easy-going, but Rupert doesn't know him that well.
| One can't be too cautious about the people one meets in Tangier. They're all weirdies of one kind or another. Me? Oh, I'm A Stranger Here Myself By MACK REYNOLDS The Place de France is the town's hub. It marks the end of Boulevard Pasteur, the main drag of the westernized part of the city, and the beginning of Rue de la Liberté, which leads down to the Grand Socco and the medina. In a three-minute walk from the Place de France you can go from an ultra-modern, California-like resort to the Baghdad of Harun al-Rashid. It's quite a town, Tangier. King-size sidewalk cafes occupy three of the strategic corners on the Place de France. The Cafe de Paris serves the best draft beer in town, gets all the better custom, and has three shoeshine boys attached to the establishment. You can sit of a sunny morning and read the Paris edition of the New York Herald Tribune while getting your shoes done up like mirrors for thirty Moroccan francs which comes to about five cents at current exchange. You can sit there, after the paper's read, sip your expresso and watch the people go by. Tangier is possibly the most cosmopolitan city in the world. In native costume you'll see Berber and Rif, Arab and Blue Man, and occasionally a Senegalese from further south. In European dress you'll see Japs and Chinese, Hindus and Turks, Levantines and Filipinos, North Americans and South Americans, and, of course, even Europeans—from both sides of the Curtain. In Tangier you'll find some of the world's poorest and some of the richest. The poorest will try to sell you anything from a shoeshine to their not very lily-white bodies, and the richest will avoid your eyes, afraid you might try to sell them something. In spite of recent changes, the town still has its unique qualities. As a result of them the permanent population includes smugglers and black-marketeers, fugitives from justice and international con men, espionage and counter-espionage agents, homosexuals, nymphomaniacs, alcoholics, drug addicts, displaced persons, ex-royalty, and subversives of every flavor. Local law limits the activities of few of these. Like I said, it's quite a town. I looked up from my Herald Tribune and said, "Hello, Paul. Anything new cooking?" He sank into the chair opposite me and looked around for the waiter. The tables were all crowded and since mine was a face he recognized, he assumed he was welcome to intrude. It was more or less standard procedure at the Cafe de Paris. It wasn't a place to go if you wanted to be alone. Paul said, "How are you, Rupert? Haven't seen you for donkey's years." The waiter came along and Paul ordered a glass of beer. Paul was an easy-going, sallow-faced little man. I vaguely remembered somebody saying he was from Liverpool and in exports. "What's in the newspaper?" he said, disinterestedly. "Pogo and Albert are going to fight a duel," I told him, "and Lil Abner is becoming a rock'n'roll singer." He grunted. "Oh," I said, "the intellectual type." I scanned the front page. "The Russkies have put up another manned satellite." "They have, eh? How big?" "Several times bigger than anything we Americans have." The beer came and looked good, so I ordered a glass too. Paul said, "What ever happened to those poxy flying saucers?" "What flying saucers?" A French girl went by with a poodle so finely clipped as to look as though it'd been shaven. The girl was in the latest from Paris. Every pore in place. We both looked after her. "You know, what everybody was seeing a few years ago. It's too bad one of these bloody manned satellites wasn't up then. Maybe they would've seen one." "That's an idea," I said. We didn't say anything else for a while and I began to wonder if I could go back to my paper without rubbing him the wrong way. I didn't know Paul very well, but, for that matter, it's comparatively seldom you ever get to know anybody very well in Tangier. Largely, cards are played close to the chest. My beer came and a plate of tapas for us both. Tapas at the Cafe de Paris are apt to be potato salad, a few anchovies, olives, and possibly some cheese. Free lunch, they used to call it in the States. Just to say something, I said, "Where do you think they came from?" And when he looked blank, I added, "The Flying Saucers." He grinned. "From Mars or Venus, or someplace." "Ummmm," I said. "Too bad none of them ever crashed, or landed on the Yale football field and said Take me to your cheerleader , or something." Paul yawned and said, "That was always the trouble with those crackpot blokes' explanations of them. If they were aliens from space, then why not show themselves?" I ate one of the potato chips. It'd been cooked in rancid olive oil. I said, "Oh, there are various answers to that one. We could probably sit around here and think of two or three that made sense." Paul was mildly interested. "Like what?" "Well, hell, suppose for instance there's this big Galactic League of civilized planets. But it's restricted, see. You're not eligible for membership until you, well, say until you've developed space flight. Then you're invited into the club. Meanwhile, they send secret missions down from time to time to keep an eye on your progress." Paul grinned at me. "I see you read the same poxy stuff I do." A Moorish girl went by dressed in a neatly tailored gray jellaba, European style high-heeled shoes, and a pinkish silk veil so transparent that you could see she wore lipstick. Very provocative, dark eyes can be over a veil. We both looked after her. I said, "Or, here's another one. Suppose you have a very advanced civilization on, say, Mars." "Not Mars. No air, and too bloody dry to support life." "Don't interrupt, please," I said with mock severity. "This is a very old civilization and as the planet began to lose its water and air, it withdrew underground. Uses hydroponics and so forth, husbands its water and air. Isn't that what we'd do, in a few million years, if Earth lost its water and air?" "I suppose so," he said. "Anyway, what about them?" "Well, they observe how man is going through a scientific boom, an industrial boom, a population boom. A boom, period. Any day now he's going to have practical space ships. Meanwhile, he's also got the H-Bomb and the way he beats the drums on both sides of the Curtain, he's not against using it, if he could get away with it." Paul said, "I got it. So they're scared and are keeping an eye on us. That's an old one. I've read that a dozen times, dished up different." I shifted my shoulders. "Well, it's one possibility." "I got a better one. How's this. There's this alien life form that's way ahead of us. Their civilization is so old that they don't have any records of when it began and how it was in the early days. They've gone beyond things like wars and depressions and revolutions, and greed for power or any of these things giving us a bad time here on Earth. They're all like scholars, get it? And some of them are pretty jolly well taken by Earth, especially the way we are right now, with all the problems, get it? Things developing so fast we don't know where we're going or how we're going to get there." I finished my beer and clapped my hands for Mouley. "How do you mean, where we're going ?" "Well, take half the countries in the world today. They're trying to industrialize, modernize, catch up with the advanced countries. Look at Egypt, and Israel, and India and China, and Yugoslavia and Brazil, and all the rest. Trying to drag themselves up to the level of the advanced countries, and all using different methods of doing it. But look at the so-called advanced countries. Up to their bottoms in problems. Juvenile delinquents, climbing crime and suicide rates, the loony-bins full of the balmy, unemployed, threat of war, spending all their money on armaments instead of things like schools. All the bloody mess of it. Why, a man from Mars would be fascinated, like." Mouley came shuffling up in his babouche slippers and we both ordered another schooner of beer. Paul said seriously, "You know, there's only one big snag in this sort of talk. I've sorted the whole thing out before, and you always come up against this brick wall. Where are they, these observers, or scholars, or spies or whatever they are? Sooner or later we'd nab one of them. You know, Scotland Yard, or the F.B.I., or Russia's secret police, or the French Sûreté, or Interpol. This world is so deep in police, counter-espionage outfits and security agents that an alien would slip up in time, no matter how much he'd been trained. Sooner or later, he'd slip up, and they'd nab him." I shook my head. "Not necessarily. The first time I ever considered this possibility, it seemed to me that such an alien would base himself in London or New York. Somewhere where he could use the libraries for research, get the daily newspapers and the magazines. Be right in the center of things. But now I don't think so. I think he'd be right here in Tangier." "Why Tangier?" "It's the one town in the world where anything goes. Nobody gives a damn about you or your affairs. For instance, I've known you a year or more now, and I haven't the slightest idea of how you make your living." "That's right," Paul admitted. "In this town you seldom even ask a man where's he's from. He can be British, a White Russian, a Basque or a Sikh and nobody could care less. Where are you from, Rupert?" "California," I told him. "No, you're not," he grinned. I was taken aback. "What do you mean?" "I felt your mind probe back a few minutes ago when I was talking about Scotland Yard or the F.B.I. possibly flushing an alien. Telepathy is a sense not trained by the humanoids. If they had it, your job—and mine—would be considerably more difficult. Let's face it, in spite of these human bodies we're disguised in, neither of us is humanoid. Where are you really from, Rupert?" "Aldebaran," I said. "How about you?" "Deneb," he told me, shaking. We had a laugh and ordered another beer. "What're you doing here on Earth?" I asked him. "Researching for one of our meat trusts. We're protein eaters. Humanoid flesh is considered quite a delicacy. How about you?" "Scouting the place for thrill tourists. My job is to go around to these backward cultures and help stir up inter-tribal, or international, conflicts—all according to how advanced they are. Then our tourists come in—well shielded, of course—and get their kicks watching it." Paul frowned. "That sort of practice could spoil an awful lot of good meat." THE END Transcriber's Note: This etext was produced from Amazing Stories December 1960. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed. Minor spelling and typographical errors have been corrected without note. | D. Paul is easy-going, but Rupert doesn't know him that well. |
How do two models cooperate to select the most confident chains? | ### Introduction
NLP tasks that require multi-hop reasoning have recently enjoyed rapid progress, especially on multi-hop question answering BIBREF0, BIBREF1, BIBREF2. Advances have benefited from rich annotations of supporting evidence, as in the popular multi-hop QA and relation extraction benchmarks, e.g., HotpotQA BIBREF3 and DocRED BIBREF4, where the evidence sentences for the reasoning process were labeled by human annotators. Such evidence annotations are crucial for modern model training, since they provide finer-grained supervision for better guiding the model learning. Furthermore, they allow a pipeline fashion of model training, with each step, such as passage ranking and answer extraction, trained as a supervised learning sub-task. This is crucial from a practical perspective, in order to reduce the memory usage when handling a large amount of inputs with advanced, large pre-trained models BIBREF5, BIBREF6, BIBREF7. Manual evidence annotation is expensive, so there are only a few benchmarks with supporting evidence annotated. Even for these datasets, the structures of the annotations are still limited, as new model designs keep emerging and they may require different forms of evidence annotations. As a result, the supervision from these datasets can still be insufficient for training accurate models. Taking question answering with multi-hop reasoning as an example, annotating only supporting passages is not sufficient to show the reasoning processes due to the lack of necessary structural information (Figure FIGREF1). One example is the order of annotated evidence, which is crucial in logic reasoning and the importance of which has also been demonstrated in text-based QA BIBREF8. The other example is how the annotated evidence pieces are connected, which requires at least the definition of arguments, such as a linking entity, concept, or event. Such information has proved useful by the recently popular entity-centric methods BIBREF9, BIBREF10, BIBREF11, BIBREF12, BIBREF0, BIBREF2 and intuitively will be a benefit to these methods if available. We propose a cooperative game approach to recovering the reasoning chains with the aforementioned necessary structural information for multi-hop QA. Each recovered chain corresponds to a list of ordered passages and each pair of adjacent passages is connected with a linking entity. Specifically, we start with a model, the Ranker, which selects a sequence of passages arriving at the answers, with the restriction that each adjacent passage pair shares at least an entity. This is essentially an unsupervised task and the selection suffers from noise and ambiguity. Therefore we introduce another model, the Reasoner, which predicts the exact linking entity that points to the next passage. The two models play a cooperative game and are rewarded when they find a consistent chain. In this way, we restrict the selection to satisfy not only the format constraints (i.e., ordered passages with connected adjacencies) but also the semantic constraints (i.e., finding the next passage given that the partial selection can be effectively modeled by a Reasoner). Therefore, the selection can be less noisy. We evaluate the proposed method on datasets with different properties, i.e., HotpotQA and MedHop BIBREF13, to cover cases with both 2-hop and 3-hop reasoning. We created labeled reasoning chains for both datasets. Experimental results demonstrate the significant advantage of our proposed approach. ### Task Definition
Reasoning Chains Examples of reasoning chains in HotpotQA and MedHop are shown in Figure FIGREF1. Formally, we aim at recovering the reasoning chain in the form of $(p_1 \rightarrow e_{1,2} \rightarrow p_2 \rightarrow e_{2,3} \rightarrow \cdots \rightarrow e_{n-1,n} \rightarrow p_n)$, where each $p_i$ is a passage and each $e_{i,i+1}$ is an entity that connects $p_i$ and $p_{i+1}$, i.e., appearing in both passages. The last passage $p_n$ in the chain contains the correct answer. We say $p_i$ connects $e_{i-1,i}$ and $e_{i,i+1}$ in the sense that it describes a relationship between the two entities. Our Task Given a QA pair $(q,a)$ and all its candidate passages $\mathcal {P}$, we can extract all possible candidate chains that satisfy the conditions mentioned above, denoted as $\mathcal {C}$. The goal of reasoning chain recovery is to extract the correct chains from all the candidates, given $q,a$ and $\mathcal {P}$ as inputs. Related Work Although there are recent interests on predicting reasoning chains for multi-hop QA BIBREF0, BIBREF14, BIBREF2, they all consider a fully supervised setting; i.e., annotated reasoning chains are available. Our work is the first to recover reasoning chains in a more general unsupervised setting, thus falling into the direction of denoising over distant supervised signals. From this perspective, the most relevant studies in the NLP field includes BIBREF15, BIBREF16 for evidence identification in open-domain QA and BIBREF17, BIBREF18, BIBREF19 for rationale recovery. ### Method
The task of recovering reasoning chains is essentially an unsupervised problem, as we have no access to annotated reasoning chains. Therefore, we resort to the noisy training signal from chains obtained by distant supervision. We first propose a conditional selection model that optimizes the passage selection by considering their orders (Section SECREF4). We then propose a cooperative Reasoner-Ranker game (Section SECREF12) in which the Reasoner recovers the linking entities that point to the next passage. This enhancement encourages the Ranker to select the chains such that their distribution is easier for a linking entity prediction model (Reasoner) to capture. Therefore, it enables our model to denoise the supervision signals while recovering chains with entity information. Figure FIGREF3 gives our overall framework, with a flow describing how the Reasoner passes additional rewards to the Ranker. ### Method ::: Passage Ranking Model
The key component of our framework is the Ranker model, which is provided with a question $q$ and $K$ passages $\mathcal {P} = \lbrace p_1, p_2 ... p_K\rbrace $ from a pool of candidates, and outputs a chain of selected passages. ### Method ::: Passage Ranking Model ::: Passage Scoring
For each step of the chain, the Ranker estimates a distribution of the selection of each passage. To this end we first encode the question and passage with a 2-layer bi-directional GRU network, resulting in an encoded question $\mathbf {Q} = \lbrace \vec{\mathbf {q}_0}, \vec{\mathbf {q}_1}, ..., \vec{\mathbf {q}_N}\rbrace $ and $\mathbf {H}_i = \lbrace \vec{\mathbf {h}_{i,0}}, \vec{\mathbf {h}_{i,1}}, ..., \vec{\mathbf {h}_{i,M_i}}\rbrace $ for each passage $p_i \in P$ of length $M_i$. Then we use the MatchLSTM model BIBREF20 to get the matching score between $\mathbf {Q}$ and each $\mathbf {H}_i$ and derive the distribution of passage selection $P(p_i|q)$ (see Appendix SECREF6 for details). We denote $P(p_i|q)=\textrm {MatchLSTM}(\mathbf {H}_i, \mathbf {Q})$ for simplicity. ### Method ::: Passage Ranking Model ::: Conditional Selection
To model passage dependency along the chain of reasoning, we use a hard selection model that builds a chain incrementally. Provided with the $K$ passages, at each step $t$ the Ranker computes $P^t(p_i|\mathbf {Q}^{t-1}), i = 0, ..., K$, which is the probability of selecting passage $p_i$ conditioned on the query and previous states representation $\mathbf {Q}^{t-1}$. Then we sample one passage $p^t_{\tau }$ according to the predicted selection probability. The first step starts with the original question $\mathbf {Q}^0$. A feed-forward network is used to project the concatenation of query encoding and selected passage encoding $\tilde{\mathbf {m}}^t_{p_{\tau }}$ back to the query space, and the new query $\mathbf {Q}^{t+1}$ is used to select the next passage. ### Method ::: Passage Ranking Model ::: Reward via Distant Supervision
We use policy gradient BIBREF21 to optimize our model. As we have no access to annotated reasoning chains during training, the reward comes from distant supervision. Specifically, we reward the Ranker if a selected passage appears as the corresponding part of a distant supervised chain in $\mathcal {C}$. The model receives immediate reward at each step of selection. In this paper we only consider chains consist of $\le 3$ passages (2-hop and 3-hop chains). For the 2-hop cases, our model predicts a chain of two passages from the candidate set $\mathcal {C}$ in the form of $p_h\rightarrow e \rightarrow p_t$. Each candidate chain satisfies that $p_t$ contains the answer, while $p_h$ and $p_t$ contain a shared entity $e$. We call $p_h$ the head passage and $p_t$ the tail passage. Let $\mathcal {P}_{T}/\mathcal {P}_{H}$ denote the set of all tail/head passages from $\mathcal {C}$. Our model receives rewards $r_h, r_t$ according to its selections: For the 3-hop cases, we need to select an additional intermediate passage $p_m$ between $p_h$ and $p_t$. If we reward any $p_m$ selection that appears in the middle of a chain in candidate chain set $\mathcal {C}$, the number of feasible options can be very large. Therefore, we make our model first select the head passage $p_h$ and the tail passage $p_t$ independently and then select $p_m$ conditioned on $(p_h,p_t)$. We further restrict that each path in $\mathcal {C}$ must have the head passage containing an entity from $q$. Then the selected $p_m$ is only rewarded if it appears in a chain in $\mathcal {C}$ that starts with $p_h$ and ends with $p_t$: ### Method ::: Cooperative Reasoner
To alleviate the noise in the distant supervision signal $\mathcal {C}$, in addition to the conditional selection, we further propose a cooperative Reasoner model, also implemented with the MatchLSTM architecture (see Appendix SECREF6), to predict the linking entity from the selected passages. Intuitively, when the Ranker makes more accurate passage selections, the Reasoner will work with less noisy data and thus is easier to succeed. Specifically, the Reasoner learns to extract the linking entity from chains selected by a well-trained Ranker, and it benefits the Ranker training by providing extra rewards. Taking 2-hop as an example, we train the Ranker and Reasoner alternatively as a cooperative game: Reasoner Step: Given the first passage $p_t$ selected by the trained Ranker, the Reasoner predicts the probability of each entity $e$ appearing in $p_t$. The Reasoner is trained with the cross-entropy loss: Ranker Step: Given the Reasoner's top-1 predicted linking entity $e$, the reward for Ranker at the $2^{\textrm {nd}}$ step is defined as: The extension to 3-hop cases is straightforward; the only difference is that the Reasoner reads both the selected $p_h$ and $p_t$ to output two entities. The Ranker receives one extra reward if the Reasoner picks the correct linking entity from $p_h$, so does $p_t$. ### Experiments ::: Settings ::: Datasets
We evaluate our path selection model on HotpotQA bridge type questions and on the MedHop dataset. In HotpotQA, the entities are pre-processed Wiki anchor link objects and in MedHop they are drug/protein database identifiers. For HotpotQA, two supporting passages are provided along with each question. We ignore the support annotations during training and use them to create ground truth on development set: following BIBREF8, we determine the order of passages according to whether a passage contains the answer. We discard ambiguous instances. For MedHop, there is no evidence annotated. Therefore we created a new evaluation dataset by manually annotating the correct paths for part of the development set: we first extract all candidate paths in form of passage triplets $(p_h, p_m, p_t)$, such that $p_h$ contains the query drug and $p_t$ contains the answer drug, and $p_h/p_m$ and $p_m/p_t$ are connected by shared proteins. We label a chain as positive if all the drug-protein or protein-protein interactions are described in the corresponding passages. Note that the positive paths are not unique for a question. During training we select chains based on the full passage set $\mathcal {P}$; at inference time we extract the chains from the candidate set $\mathcal {C}$ (see Section SECREF2). ### Experiments ::: Settings ::: Baselines and Evaluation Metric
We compare our model with (1) random baseline, which randomly selects a candidate chain from the distant supervision chain set $\mathcal {C}$; and (2) distant supervised MatchLSTM, which uses the same base model as ours but scores and selects the passages independently. We use accuracy as our evaluation metric. As HotpotQA does not provide ground-truth linking entities, we only evaluate whether the supporting passages are fully recovered (yet our model still output the full chains). For MedHop we evaluate whether the whole predicted chain is correct. More details can be found in Appendix SECREF7. We use BIBREF24 as word embedding for HotpotQA, and BIBREF25 for MedHop. ### Experiments ::: Results ::: HotpotQA
We first evaluate on the 2-hop HotpotQA task. Our best performed model first selects the tail passage $p_t$ and then the head passage $p_h$, because the number of candidates of tail is smaller ($\sim $2 per question). Table TABREF21 shows the results. First, training a ranker with distant supervision performs significantly better than the random baseline, showing that the training process itself has a certain degree of denoising ability to distinguish the more informative signals from distant supervision labels. By introducing additional inductive bias of orders, the conditional selection model further improves with a large margin. Finally, our cooperative game gives the best performance, showing that a trained Reasoner has the ability of ignoring entity links that are irrelevant to the reasoning chain. Table TABREF22 demonstrates the effect of selecting directions, together with the methods' recall on head passages and tail passages. The latter is evaluated on a subset of bridge-type questions in HotpotQA which has no ambiguous support annotations in passage orders; i.e., among the two human-labeled supporting passages, only one contains the answer and thus must be a tail. The results show that selecting tail first performs better. The cooperative game mainly improves the head selection. ### Experiments ::: Results ::: MedHop
Results in table TABREF21 show that recovering chains from MedHop is a much harder task: first, the large number of distant supervision chains in $\mathcal {C}$ introduce too much noise so the Distant Supervised Ranker improves only 3%; second, the dependent model leads to no improvement because $\mathcal {C}$ is strictly ordered given our data construction. Our cooperative game manages to remain effective and gives further improvement. ### Conclusions
In this paper we propose the problem of recovering reasoning chains in multi-hop QA from weak supervision signals. Our model adopts an cooperative game approach where a ranker and a reasoner cooperate to select the most confident chains. Experiments on the HotpotQA and MedHop benchmarks show the effectiveness of the proposed approach. ### Details of MatchLSTMs for Passage Scoring and Reasoner ::: MatchLSTM for Passage Scoring
Given the embeddings $\mathbf {Q} = \lbrace \vec{\mathbf {q}_0}, \vec{\mathbf {q}_1}, ..., \vec{\mathbf {q}_N}\rbrace $ of the question $q$, and $\mathbf {H}_i = \lbrace \vec{\mathbf {h}_{i,0}}, \vec{\mathbf {h}_{i,1}}, ..., \vec{\mathbf {h}_{i,M_i}}\rbrace $ of each passage $p_i \in P$, we use the MatchLSTM BIBREF20 to match $\mathbf {Q}$ and $\mathbf {H}_i$ as follows: The final vector $\tilde{\mathbf {m}}_i$ represents the matching state between $q$ and $p_i$. All the $\tilde{\mathbf {m}}_i$s are then passed to a linear layer that outputs the ranking score of each passage. We apply softmax over the scores to get the probability of passage selection $P(p_i|q)$. We denote the above computation as $P(p_i|q)=\textrm {MatchLSTM}(\mathbf {H}_i, \mathbf {Q})$ for simplicity. ### Details of MatchLSTMs for Passage Scoring and Reasoner ::: MatchLSTM for Reasoner
Given the question embedding $\mathbf {Q}^r = \lbrace \vec{\mathbf {q}^r_0}, \vec{\mathbf {q}^r_1}, ..., \vec{\mathbf {q}^r_N}\rbrace $ and the input passage embedding $\mathbf {H}^r = \lbrace \vec{\mathbf {h}^r_{0}}, \vec{\mathbf {h}^r_{1}}, ..., \vec{\mathbf {h}^r_{M}}\rbrace $ of $p$, the Reasoner predicts the probability of each entity in the passage being the linking entity of the next passage in the chain. We use a reader model similar to BIBREF3 as our Reasoner network. We first describe an attention sub-module. Given input sequence embedding $\mathbf {A} = \lbrace \vec{\mathbf {a}_0}, \vec{\mathbf {a}_1}, ..., \vec{\mathbf {a}_N}\rbrace $ and $\mathbf {B} = \lbrace \vec{\mathbf {b}_{0}}, \vec{\mathbf {b}_{1}}, ..., \vec{\mathbf {b}_{M}}\rbrace $, we define $\tilde{\mathcal {M}} = \text{Attention}(\mathbf {A}, \mathbf {B})$: where FFN denotes a feed forward layer which projects the concatenated embedding back to the original space. The Reasoner network consists of multiple attention layers, together with a bidirectional GRU encoder and skip connection. For each token $e_k, k = 0, 1,..., M$ represented by $h^r_{p,k}$ at the corresponding location, we have: where $g$ is the classification layer, softmax is applied across all entities to get the probability. We denote the computation above as $P^r(e_k| \mathbf {p}) = \textrm {MatchLSTM.Reader}(e_k, \mathbf {p})$ for simplicity. ### Definition of Chain Accuracy
In HotpotQA, on average we can find 6 candidate chains (2-hop) in a instance, and the human labeled true reasoning chain is unique. A predicted chain is correct if the chain only contains all supporting passages (exact match of passages). In MedHop, on average we can find 30 candidate chains (3-hop). For each candidate chain our human annotators labeled whether it is correct or not, and the correct reasoning chain is not unique. A predicted chain is correct if it is one of the chains that human labeled as correct. The accuracy is defined as the ratio: Figure 1: An example of reasoning chains in HotpotQA (2- hop) and MedHop (3-hop). HotpotQA provides only supporting passages {P3, P9}, without order and linking information. Figure 2: Model overview. The cooperative Ranker and Reasoner are trained alternatively. The Ranker selects a passage p at each step conditioned on the question q and history selection, and receives reward r1 if p is evidence. Conditioned on q, the Reasoner predicts which entity from p links to the next evidence passage. The Ranker receives extra reward r2 if its next selection is connected by the entity predicted by the Reasoner. Both q and answer a are model inputs. While q is fed to the Ranker/Reasoner as input, empirically the best way of using a is for constructing the candidate set thus computing the reward r1. We omit the flow from q/a for simplicity. Table 1: Reasoning Chain selection results. Table 2: Ablation test on HotpotQA. | Reasoner learns to extract the linking entity from chains selected by a well-trained Ranker, and it benefits the Ranker training by providing extra rewards |
Of the following options, which best describe Meredith?
A. Bold and pretty
B. Brave and adventurous
C. Beautiful and brave
D. Smart and kindhearted
| BIG ANCESTOR By F. L. WALLACE Illustrated by EMSH [Transcriber's Note: This etext was produced from Galaxy Science Fiction November 1954. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed.] Man's family tree was awesome enough to give every galactic race an inferiority complex—but then he tried to climb it! In repose, Taphetta the Ribboneer resembled a fancy giant bow on a package. His four flat legs looped out and in, the ends tucked under his wide, thin body, which constituted the knot at the middle. His neck was flat, too, arching out in another loop. Of all his features, only his head had appreciable thickness and it was crowned with a dozen long though narrower ribbons. Taphetta rattled the head fronds together in a surprisingly good imitation of speech. "Yes, I've heard the legend." "It's more than a legend," said Sam Halden, biologist. The reaction was not unexpected—non-humans tended to dismiss the data as convenient speculation and nothing more. "There are at least a hundred kinds of humans, each supposedly originating in strict seclusion on as many widely scattered planets. Obviously there was no contact throughout the ages before space travel— and yet each planetary race can interbreed with a minimum of ten others ! That's more than a legend—one hell of a lot more!" "It is impressive," admitted Taphetta. "But I find it mildly distasteful to consider mating with someone who does not belong to my species." "That's because you're unique," said Halden. "Outside of your own world, there's nothing like your species, except superficially, and that's true of all other creatures, intelligent or not, with the sole exception of mankind. Actually, the four of us here, though it's accidental, very nearly represent the biological spectrum of human development. "Emmer, a Neanderthal type and our archeologist, is around the beginning of the scale. I'm from Earth, near the middle, though on Emmer's side. Meredith, linguist, is on the other side of the middle. And beyond her, toward the far end, is Kelburn, mathematician. There's a corresponding span of fertility. Emmer just misses being able to breed with my kind, but there's a fair chance that I'd be fertile with Meredith and a similar though lesser chance that her fertility may extend to Kelburn." Taphetta rustled his speech ribbons quizzically. "But I thought it was proved that some humans did originate on one planet, that there was an unbroken line of evolution that could be traced back a billion years." "You're thinking of Earth," said Halden. "Humans require a certain kind of planet. It's reasonable to assume that, if men were set down on a hundred such worlds, they'd seem to fit in with native life-forms on a few of them. That's what happened on Earth; when Man arrived, there was actually a manlike creature there. Naturally our early evolutionists stretched their theories to cover the facts they had. "But there are other worlds in which humans who were there before the Stone Age aren't related to anything else there. We have to conclude that Man didn't originate on any of the planets on which he is now found. Instead, he evolved elsewhere and later was scattered throughout this section of the Milky Way." "And so, to account for the unique race that can interbreed across thousands of light-years, you've brought in the big ancestor," commented Taphetta dryly. "It seems an unnecessary simplification." "Can you think of a better explanation?" asked Kelburn. "Something had to distribute one species so widely and it's not the result of parallel evolution—not when a hundred human races are involved, and only the human race." "I can't think of a better explanation." Taphetta rearranged his ribbons. "Frankly, no one else is much interested in Man's theories about himself." It was easy to understand the attitude. Man was the most numerous though not always the most advanced—Ribboneers had a civilization as high as anything in the known section of the Milky Way, and there were others—and humans were more than a little feared. If they ever got together—but they hadn't except in agreement as to their common origin. Still, Taphetta the Ribboneer was an experienced pilot and could be very useful. A clear statement of their position was essential in helping him make up his mind. "You've heard of the adjacency mating principle?" asked Sam Halden. "Vaguely. Most people have if they've been around men." "We've got new data and are able to interpret it better. The theory is that humans who can mate with each other were once physically close. We've got a list of all our races arranged in sequence. If planetary race F can mate with race E back to A and forward to M, and race G is fertile only back to B, but forward to O, then we assume that whatever their positions are now, at once time G was actually adjacent to F, but was a little further along. When we project back into time those star systems on which humans existed prior to space travel, we get a certain pattern. Kelburn can explain it to you." The normally pink body of the Ribboneer flushed slightly. The color change was almost imperceptible, but it was enough to indicate that he was interested. Kelburn went to the projector. "It would be easier if we knew all the stars in the Milky Way, but though we've explored only a small portion of it, we can reconstruct a fairly accurate representation of the past." He pressed the controls and stars twinkled on the screen. "We're looking down on the plane of the Galaxy. This is one arm of it as it is today and here are the human systems." He pressed another control and, for purposes of identification, certain stars became more brilliant. There was no pattern, merely a scattering of stars. "The whole Milky Way is rotating. And while stars in a given region tend to remain together, there's also a random motion. Here's what happens when we calculate the positions of stars in the past." Flecks of light shifted and flowed across the screen. Kelburn stopped the motion. "Two hundred thousand years ago," he said. There was a pattern of the identified stars. They were spaced at fairly equal intervals along a regular curve, a horseshoe loop that didn't close, though if the ends were extended, the lines would have crossed. Taphetta rustled. "The math is accurate?" "As accurate as it can be with a million-plus body problem." "And that's the hypothetical route of the unknown ancestor?" "To the best of our knowledge," said Kelburn. "And whereas there are humans who are relatively near and not fertile, they can always mate with those they were adjacent to two hundred thousand years ago !" "The adjacency mating principle. I've never seen it demonstrated," murmured Taphetta, flexing his ribbons. "Is that the only era that satisfies the calculations?" "Plus or minus a hundred thousand years, we can still get something that might be the path of a spaceship attempting to cover a representative section of territory," said Kelburn. "However, we have other ways of dating it. On some worlds on which there are no other mammals, we're able to place the first human fossils chronologically. The evidence is sometimes contradictory, but we believe we've got the time right." Taphetta waved a ribbon at the chart. "And you think that where the two ends of the curve cross is your original home?" "We think so," said Kelburn. "We've narrowed it down to several cubic light-years—then. Now it's far more. And, of course, if it were a fast-moving star, it might be completely out of the field of our exploration. But we're certain we've got a good chance of finding it this trip." "It seems I must decide quickly." The Ribboneer glanced out the visionport, where another ship hung motionless in space beside them. "Do you mind if I ask other questions?" "Go ahead," Kelburn invited sardonically. "But if it's not math, you'd better ask Halden. He's the leader of the expedition." Halden flushed; the sarcasm wasn't necessary. It was true that Kelburn was the most advanced human type present, but while there were differences, biological and in the scale of intelligence, it wasn't as great as once was thought. Anyway, non-humans weren't trained in the fine distinctions that men made among themselves. And, higher or lower, he was as good a biologist as the other was a mathematician. And there was the matter of training; he'd been on several expeditions and this was Kelburn's first trip. Damn it, he thought, that rated some respect. The Ribboneer shifted his attention. "Aside from the sudden illness of your pilot, why did you ask for me?" "We didn't. The man became sick and required treatment we can't give him. Luckily, a ship was passing and we hailed it because it's four months to the nearest planet. They consented to take him back and told us that there was a passenger on board who was an experienced pilot. We have men who could do the job in a makeshift fashion, but the region we're heading for, while mapped, is largely unknown. We'd prefer to have an expert—and Ribboneers are famous for their navigational ability." Taphetta crinkled politely at the reference to his skill. "I had other plans, but I can't evade professional obligations, and an emergency such as this should cancel out any previous agreements. Still, what are the incentives?" Sam Halden coughed. "The usual, plus a little extra. We've copied the Ribboneer's standard nature, simplifying it a little and adding a per cent here and there for the crew pilot and scientist's share of the profits from any discoveries we may make." "I'm complimented that you like our contract so well," said Taphetta, "but I really must have our own unsimplified version. If you want me, you'll take my contract. I came prepared." He extended a tightly bound roll that he had kept somewhere on his person. They glanced at one another as Halden took it. "You can read it if you want," offered Taphetta. "But it will take you all day—it's micro-printing. However, you needn't be afraid that I'm defrauding you. It's honored everywhere we go and we go nearly everywhere in this sector—places men have never been." There was no choice if they wanted him, and they did. Besides, the integrity of Ribboneers was not to be questioned. Halden signed. "Good." Taphetta crinkled. "Send it to the ship; they'll forward it for me. And you can tell the ship to go on without me." He rubbed his ribbons together. "Now if you'll get me the charts, I'll examine the region toward which we're heading." Firmon of hydroponics slouched in, a tall man with scanty hair and an equal lack of grace. He seemed to have difficulty in taking his eyes off Meredith, though, since he was a notch or so above her in the mating scale, he shouldn't have been so interested. But his planet had been inexplicably slow in developing and he wasn't completely aware of his place in the human hierarchy. Disdainfully, Meredith adjusted a skirt that, a few inches shorter, wouldn't have been a skirt at all, revealing, while doing so, just how long and beautiful a woman's legs could be. Her people had never given much thought to physical modesty and, with legs like that, it was easy to see why. Muttering something about primitive women, Firmon turned to the biologist. "The pilot doesn't like our air." "Then change it to suit him. He's in charge of the ship and knows more about these things than I do." "More than a man?" Firmon leered at Meredith and, when she failed to smile, added plaintively, "I did try to change it, but he still complains." Halden took a deep breath. "Seems all right to me." "To everybody else, too, but the tapeworm hasn't got lungs. He breathes through a million tubes scattered over his body." It would do no good to explain that Taphetta wasn't a worm, that his evolution had taken a different course, but that he was in no sense less complex than Man. It was a paradox that some biologically higher humans hadn't developed as much as lower races and actually weren't prepared for the multitude of life-forms they'd meet in space. Firmon's reaction was quite typical. "If he asks for cleaner air, it's because his system needs it," said Halden. "Do anything you can to give it to him." "Can't. This is as good as I can get it. Taphetta thought you could do something about it." "Hydroponics is your job. There's nothing I can do." Halden paused thoughtfully. "Is there something wrong with the plants?" "In a way, I guess, and yet not really." "What is it, some kind of toxic condition?" "The plants are healthy enough, but something's chewing them down as fast as they grow." "Insects? There shouldn't be any, but if there are, we've got sprays. Use them." "It's an animal," said Firmon. "We tried poison and got a few, but now they won't touch the stuff. I had electronics rig up some traps. The animals seem to know what they are and we've never caught one that way." Halden glowered at the man. "How long has this been going on?" "About three months. It's not bad; we can keep up with them." It was probably nothing to become alarmed at, but an animal on the ship was a nuisance, doubly so because of their pilot. "Tell me what you know about it," said Halden. "They're little things." Firmon held out his hands to show how small. "I don't know how they got on, but once they did, there were plenty of places to hide." He looked up defensively. "This is an old ship with new equipment and they hide under the machinery. There's nothing we can do except rebuild the ship from the hull inward." Firmon was right. The new equipment had been installed in any place just to get it in and now there were inaccessible corners and crevices everywhere that couldn't be closed off without rebuilding. They couldn't set up a continuous watch and shoot the animals down because there weren't that many men to spare. Besides, the use of weapons in hydroponics would cause more damage to the thing they were trying to protect than to the pest. He'd have to devise other ways. Sam Halden got up. "I'll take a look and see what I can do." "I'll come along and help," said Meredith, untwining her legs and leaning against him. "Your mistress ought to have some sort of privileges." Halden started. So she knew that the crew was calling her that! Perhaps it was intended to discourage Firmon, but he wished she hadn't said it. It didn't help the situation at all. Taphetta sat in a chair designed for humans. With a less flexible body, he wouldn't have fitted. Maybe it wasn't sitting, but his flat legs were folded neatly around the arms and his head rested comfortably on the seat. The head ribbons, which were his hands and voice, were never quite still. He looked from Halden to Emmer and back again. "The hydroponics tech tells me you're contemplating an experiment. I don't like it." Halden shrugged. "We've got to have better air. It might work." "Pests on the ship? It's filthy! My people would never tolerate it!" "Neither do we." The Ribboneer's distaste subsided. "What kind of creatures are they?" "I have a description, though I've never seen one. It's a small four-legged animal with two antennae at the lower base of its skull. A typical pest." Taphetta rustled. "Have you found out how it got on?" "It was probably brought in with the supplies," said the biologist. "Considering how far we've come, it may have been any one of a half a dozen planets. Anyway, it hid, and since most of the places it had access to were near the outer hull, it got an extra dose of hard radiation, or it may have nested near the atomic engines; both are possibilities. Either way, it mutated, became a different animal. It's developed a tolerance for the poisons we spray on plants. Other things it detects and avoids, even electronic traps." "Then you believe it changed mentally as well as physically, that it's smarter?" "I'd say that, yes. It must be a fairly intelligent creature to be so hard to get rid of. But it can be lured into traps, if the bait's strong enough." "That's what I don't like," said Taphetta, curling. "Let me think it over while I ask questions." He turned to Emmer. "I'm curious about humans. Is there anything else you can tell me about the hypothetical ancestor?" Emmer didn't look like the genius he was—a Neanderthal genius, but nonetheless a real one. In his field, he rated very high. He raised a stubble-flecked cheek from a large thick-fingered paw and ran shaggy hands through shaggier hair. "I can speak with some authority," he rumbled. "I was born on a world with the most extensive relics. As a child, I played in the ruins of their camp." "I don't question your authority," crinkled Taphetta. "To me, all humans—late or early and male or female—look remarkably alike. If you are an archeologist, that's enough for me." He paused and flicked his speech ribbons. "Camp, did you say?" Emmer smiled, unsheathing great teeth. "You've never seen any pictures? Impressive, but just a camp, monolithic one-story structures, and we'd give something to know what they're made of. Presumably my world was one of the first they stopped at. They weren't used to roughing it, so they built more elaborately than they did later on. One-story structures and that's how we can guess at their size. The doorways were forty feet high." "Very large," agreed Taphetta. It was difficult to tell whether he was impressed. "What did you find in the ruins?" "Nothing," said Emmer. "There were buildings there and that was all, not a scrap of writing or a tool or a single picture. They covered a route estimated at thirty thousand light-years in less than five thousand years—and not one of them died that we have a record of." "A faster-than-light drive and an extremely long life," mused Taphetta. "But they didn't leave any information for their descendants. Why?" "Who knows? Their mental processes were certainly far different from ours. They may have thought we'd be better off without it. We do know they were looking for a special kind of planet, like Earth, because they visited so many of that type, yet different from it because they never stayed. They were pretty special people themselves, big and long-lived, and maybe they couldn't survive on any planet they found. Perhaps they had ways of determining there wasn't the kind of planet they needed in the entire Milky Way. Their science was tremendously advanced and when they learned that, they may have altered their germ plasm and left us, hoping that some of us would survive. Most of us did." "This special planet sounds strange," murmured Taphetta. "Not really," said Emmer. "Fifty human races reached space travel independently and those who did were scattered equally among early and late species. It's well known that individuals among my people are often as bright as any of Halden's or Meredith's, but as a whole we don't have the total capacity that later Man does, and yet we're as advanced in civilization. The difference? It must lie somewhere in the planets we live on and it's hard to say just what it is." "What happened to those who didn't develop space travel?" asked Taphetta. "We helped them," said Emmer. And they had, no matter who or what they were, biologically late or early, in the depths of the bronze age or the threshold of atomic—because they were human. That was sometimes a frightening thing for non-humans, that the race stuck together. They weren't actually aggressive, but their total number was great and they held themselves aloof. The unknown ancestor again. Who else had such an origin and, it was tacitly assumed, such a destiny? Taphetta changed his questioning. "What do you expect to gain from this discovery of the unknown ancestor?" It was Halden who answered him. "There's the satisfaction of knowing where we came from." "Of course," rustled the Ribboneer. "But a lot of money and equipment was required for this expedition. I can't believe that the educational institutions that are backing you did so purely out of intellectual curiosity." "Cultural discoveries," rumbled Emmer. "How did our ancestors live? When a creature is greatly reduced in size, as we are, more than physiology is changed—the pattern of life itself is altered. Things that were easy for them are impossible for us. Look at their life span." "No doubt," said Taphetta. "An archeologist would be interested in cultural discoveries." "Two hundred thousand years ago, they had an extremely advanced civilization," added Halden. "A faster-than-light drive, and we've achieved that only within the last thousand years." "But I think we have a better one than they did," said the Ribboneer. "There may be things we can learn from them in mechanics or physics, but wouldn't you say they were better biologists than anything else?" Halden nodded. "Agreed. They couldn't find a suitable planet. So, working directly with their germ plasm, they modified themselves and produced us. They were master biologists." "I thought so," said Taphetta. "I never paid much attention to your fantastic theories before I signed to pilot this ship, but you've built up a convincing case." He raised his head, speech ribbons curling fractionally and ceaselessly. "I don't like to, but we'll have to risk using bait for your pest." He'd have done it anyway, but it was better to have the pilot's consent. And there was one question Halden wanted to ask; it had been bothering him vaguely. "What's the difference between the Ribboneer contract and the one we offered you? Our terms are more liberal." "To the individual, they are, but it won't matter if you discover as much as you think you will. The difference is this: My terms don't permit you to withhold any discovery for the benefit of one race." Taphetta was wrong; there had been no intention of withholding anything. Halden examined his own attitudes. He hadn't intended, but could he say that was true of the institutions backing the expedition? He couldn't, and it was too late now—whatever knowledge they acquired would have to be shared. That was what Taphetta had been afraid of—there was one kind of technical advancement that multiplied unceasingly. The race that could improve itself through scientific control of its germ plasm had a start that could never be headed. The Ribboneer needn't worry now. "Why do we have to watch it on the screen?" asked Meredith, glancing up. "I'd rather be in hydroponics." Halden shrugged. "They may or may not be smarter than planetbound animals, but they're warier. They don't come out when anyone's near." Lights dimmed in the distant hydroponic section and the screen with it, until he adjusted the infra-red frequencies. He motioned to the two crew members, each with his own peculiar screen, below which was a miniature keyboard. "Ready?" When they nodded, Halden said: "Do as you've rehearsed. Keep noise at a minimum, but when you do use it, be vague. Don't try to imitate them exactly." At first, nothing happened on the big screen, and then a gray shape crept out. It slid through leaves, listened intently before coming forward. It jumped off one hydroponic section and fled across the open floor to the next. It paused, eyes glittering and antennae twitching. Looking around once, it leaped up, seizing the ledge and clawing up the side of the tank. Standing on top and rising to its haunches, it began nibbling what it could reach. Suddenly it whirled. Behind it and hitherto unnoticed was another shape, like it but larger. The newcomer inched forward. The small one retreated, skittering nervously. Without warning, the big one leaped and the small one tried to flee. In a few jumps, the big one caught up and mauled the other unmercifully. It continued to bite even after the little one lay still. At last it backed off and waited, watching for signs of motion. There was none. Then it turned to the plant. When it had chewed off everything within reach, it climbed into the branches. The little one twitched, moved a leg, and cautiously began dragging itself away. It rolled off the raised section and surprisingly made no noise as it fell. It seemed to revive, shaking itself and scurrying away, still within range of the screen. Against the wall was a small platform. The little one climbed on top and there found something that seemed to interest it. It sniffed around and reached and felt the discovery. Wounds were forgotten as it snatched up the object and frisked back to the scene of its recent defeat. This time it had no trouble with the raised section. It leaped and landed on top and made considerable noise in doing so. The big animal heard and twisted around. It saw and clambered down hastily, jumping the last few feet. Squealing, it hit the floor and charged. The small one stood still till the last instant—and then a paw flickered out and an inch-long knife blade plunged into the throat of the charging creature. Red spurted out as the bigger beast screamed. The knife flashed in and out until the big animal collapsed and stopped moving. The small creature removed the knife and wiped it on the pelt of its foe. Then it scampered back to the platform on which the knife had been found— and laid it down . At Halden's signal, the lights flared up and the screen became too bright for anything to be visible. "Go in and get them," said Halden. "We don't want the pests to find out that the bodies aren't flesh." "It was realistic enough," said Meredith as the crewmen shut off their machines and went out. "Do you think it will work?" "It might. We had an audience." "Did we? I didn't notice." Meredith leaned back. "Were the puppets exactly like the pests? And if not, will the pests be fooled?" "The electronic puppets were a good imitation, but the animals don't have to identify them as their species. If they're smart enough, they'll know the value of a knife, no matter who uses it." "What if they're smarter? Suppose they know a knife can't be used by a creature without real hands?" "That's part of our precautions. They'll never know until they try—and they'll never get away from the trap to try." "Very good. I never thought of that," said Meredith, coming closer. "I like the way your primitive mind works. At times I actually think of marrying you." "Primitive," he said, alternately frozen and thawed, though he knew that, in relation to her, he was not advanced. "It's almost a curse, isn't it?" She laughed and took the curse away by leaning provocatively against him. "But barbaric lovers are often nice." Here we go again, he thought drearily, sliding his arm around her. To her, I'm merely a passionate savage. They went to his cabin. She sat down, smiling. Was she pretty? Maybe. For her own race, she wasn't tall, only by Terran standards. Her legs were disproportionately long and well shaped and her face was somewhat bland and featureless, except for a thin, straight, short nose. It was her eyes that made the difference, he decided. A notch or two up the scale of visual development, her eyes were larger and she could see an extra color on the violet end of the spectrum. She settled back and looked at him. "It might be fun living with you on primeval Earth." He said nothing; she knew as well as he that Earth was as advanced as her own world. She had something else in mind. "I don't think I will, though. We might have children." "Would it be wrong?" he asked. "I'm as intelligent as you. We wouldn't have subhuman monsters." "It would be a step up—for you." Under her calm, there was tension. It had been there as long as he'd known her, but it was closer to the surface now. "Do I have the right to condemn the unborn? Should I make them start lower than I am?" The conflict was not new nor confined to them. In one form or another, it governed personal relations between races that were united against non-humans, but held sharp distinctions themselves. "I haven't asked you to marry me," he said bluntly. "Because you're afraid I'd refuse." It was true; no one asked a member of a higher race to enter a permanent union. "Why did you ever have anything to do with me?" demanded Halden. "Love," she said gloomily. "Physical attraction. But I can't let it lead me astray." "Why not make a play for Kelburn? If you're going to be scientific about it, he'd give you children of the higher type." "Kelburn." It didn't sound like a name, the way she said it. "I don't like him and he wouldn't marry me." "He wouldn't, but he'd give you children if you were humble enough. There's a fifty per cent chance you might conceive." She provocatively arched her back. Not even the women of Kelburn's race had a body like hers and she knew it. "Racially, there should be a chance," she said. "Actually, Kelburn and I would be infertile." "Can you be sure?" he asked, knowing it was a poor attempt to act unconcerned. "How can anyone be sure on a theoretical basis?" she asked, an oblique smile narrowing her eyes. "I know we can't." His face felt anesthetized. "Did you have to tell me that?" She got up and came to him. She nuzzled against him and his reaction was purely reflexive. His hand swung out and he could feel the flesh give when his knuckles struck it. She fell back and dazedly covered her face with her hand. When she took it away, blood spurted. She groped toward the mirror and stood in front of it. She wiped the blood off, examining her features carefully. "You've broken my nose," she said factually. "I'll have to stop the blood and pain." She pushed her nose back into place and waggled it to make sure. She closed her eyes and stood silent and motionless. Then she stepped back and looked at herself critically. "It's set and partially knitted. I'll concentrate tonight and have it healed by morning." She felt in the cabinet and attached an invisible strip firmly across the bridge. Then she came over to him. "I wondered what you'd do. You didn't disappoint me." He scowled miserably at her. Her face was almost plain and the bandage, invisible or not, didn't improve her appearance any. How could he still feel that attraction to her? "Try Emmer," he suggested tiredly. "He'll find you irresistible, and he's even more savage than I am." "Is he?" She smiled enigmatically. "Maybe, in a biological sense. Too much, though. You're just right." He sat down on the bed. Again there was only one way of knowing what Emmer would do—and she knew. She had no concept of love outside of the physical, to make use of her body so as to gain an advantage—what advantage?—for the children she intended to have. Outside of that, nothing mattered, and for the sake of alloying the lower with the higher, she was as cruel to herself as she was to him. And yet he wanted her. "I do think I love you," she said. "And if love's enough, I may marry you in spite of everything. But you'll have to watch out whose children I have." She wriggled into his arms. The racial disparity was great and she had provoked him, but it was not completely her fault. Besides.... Besides what? She had a beautiful body that could bear superior children—and they might be his. He twisted away. With those thoughts, he was as bad as she was. Were they all that way, every one of them, crawling upward out of the slime toward the highest goal they could conceive of? Climbing over—no, through —everybody they could coerce, seduce or marry—onward and upward. He raised his hand, but it was against himself that his anger was turned. "Careful of the nose," she said, pressing against him. "You've already broken it once." He kissed her with sudden passion that even he knew was primitive. | A. Bold and pretty |
As per the histology report dated 03/18/2017, what was the mitotic figure count per 10 high-power fields for Mr. Havers' tumor?
Choose the correct answer from the following options:
A. 0 mitotic figures
B. 1 mitotic figure
C. 2 mitotic figures
D. 3 mitotic figures
E. 4 mitotic figures
| ### Patient Report 0
**Dear colleague, **
We are reporting on our patient, John Havers, born on 05/29/1953, who
received an MRI of the right proximal thigh for further clarification of
a potential tumor.
**MRI of the right thigh, plain and with contrast agent, on
02/19/2017:**
[Technique]{.underline}: Surface coil, localization scan, coronal T1 SE,
transverse, coronal, sagittal T2 TSE with fat suppression. After
intravenous contrast administration, T1-TSE transverse and T1-TSE FS
(coronal, T2 TSE FS coronal as an additional fat-saturated sequence in
the same section level for exploring relevant edema).
[Findings]{.underline}: Normal bone marrow signal consistent with age.
No signs of fractures. Coexistence of moderate degenerative changes in
the hip joints, more pronounced on the right than on the left. Mild
activation of the muscles in the left proximal adductor region. Ventral
to the gracilis muscle and dorsal to the sartorius muscle at the level
of the middle third of the right thigh is a subfascial intermuscular
oval mass lesion with a high-signal appearance on T2-weighted images and
a low-signal appearance on T1-weighted images. It is partially septated,
well-demarcated, and shows strong contrast enhancement. No evidence of
blood degradation products. Dimensions are 35 x 45 x 40 mm. No evidence
of suspiciously enlarged lymph nodes. Other assessed soft tissues are
unremarkable for the patient\'s age.
[Assessment]{.underline}: Overall, a high suspicion of a mucinous mass
lesion in the region of the right adductor compartment. Differential
Diagnosis: Mucinous liposarcoma. Further histological evaluation is
strongly recommended.
**Current Recommendations:** Presentation at the clinic for surgery for
further differential diagnostic clarification.
### Patient Report 1
**Dear colleague, **
We are reporting on our patient, John Havers, born on 05/29/1953. He was
under our inpatient care from 03/10/2017 to 03/12/2017.
**Diagnosis:** Soft tissue tumor of the right proximal thigh
**Other Diagnoses:**
- Arterial hypertension
- Non-insulin-dependent diabetes mellitus
- Coronary artery disease with stent placement
- Nicotine abuse (80-100 pack-years)
- Arterial hypertension
- Status post apoplexy
- Status post cataract surgery
- Status post right hip total hip replacement
- Status post Polypectomy for polyposis coli (minimal dysplasia)
- Status post appendectomy
- M. Meniere
**Allergies**: Hay fever
**Treatment**: Incisional biopsy on 03/10/2017
**Histology:**
[Microscopy]{.underline}: (Hematoxylin and Eosin staining):
Histologically, an infiltrate of a mesenchymal neoplasm is evident in a
section prepared by us and stained with HE. There are areas with an
estimated tumor percentage of approximately 90% that were selected and
labeled for molecular pathology analysis.
[Molecular Pathology]{.underline}: After macrodissection of labeled
tumor areas from unstained consecutive sections, RNA was extracted and
analyzed using focused next-generation sequencing technology. The
analysis was performed using FusioPlex Sarcoma v2 assays, allowing
detection of fusions in 63 genes.
**Medical History:** We may kindly assume that you are familiar with Mr.
Havers's medical history. The patient presented to our surgery clinic
due to a mass in the right proximal thigh. The swelling was first
noticed approximately 3 months ago and has shown significant enlargement
since. The patient subsequently consulted a general surgeon, who
referred him to our center after performing an MRI, suspecting an
intramuscular liposarcoma. After presenting the case to our
interdisciplinary tumor board, the decision was made to perform an
incisional biopsy. The patient was admitted for the above procedure on
03/10/2017.
**Physical Examination:** On clinical examination, a patient in slightly
reduced general and nutritional status was observed. Approximately 6 x 7
x 4 cm-sized tumor in the right proximal thigh, well mobile,
intramuscular. Numbness in both legs at L5/S1.
No change in skin color. No fluctuation or redness. The rest of the
clinical examination was unremarkable.
**Treatment and Progression:** Following routine preoperative
preparations and informed consent, the above-mentioned procedure was
performed under general anesthesia on 03/10/2017. The intraoperative and
postoperative courses were uncomplicated.
Initial mild swelling regressed over time. The inserted drainage was
removed on the second postoperative day. The patient mobilized
independently on the ward. Pain management was provided as needed.
With the patient\'s subjective well-being and inconspicuous wound
conditions, we were able to discharge Mr. Havers on 03/12/2017 for
outpatient follow-up.
**Current Recommendations:**
- Suture material to be shortened on the 14th postoperative day.
- Follow-up appointments in our outpatient clinic
**Medication upon Discharge:**
**Medication** **Dosage** **Frequency**
-------------------------------------- ------------ ---------------
Empagliflozin (Jardiance) 10 mg 1-0-0-0
Metformin Hydrochloride (Glucophage) 1000 mg 1-0-1-0
Atorvastatin Calcium (Lipitor) 21.7 mg 0-0-1-0
Metoprolol Tartrate (Lopressor) 50 mg 0.5-0-0.5-0
Aspirin 100 mg 1-0-0-0
Pantoprazole Sodium (Protonix) 22.6 mg 1-0-0-0
**Lab results upon Discharge: **
**Parameter** **Results** **Reference Range**
---------------------- ------------- ---------------------
Sodium 138 mEq/L 136-145 mEq/L
Potassium 4.9 mEq/L 3.5-4.5 mEq/L
Creatinine 0.81 mg/dL 0.70-1.20 mg/dL
Estimated GFR \- \-
Urea 38 mg/dL 17-48 mg/dL
C-Reactive Protein 2.6 mg/dL \< 5.0 mg/dL
Complete Blood Count \- \-
Hemoglobin 16.7 g/dL 13.5-17.0 g/dL
Hematocrit 49.5% 39.5-50.5%
Erythrocytes 5.2 M/µL 4.3-5.8 M/µL
Leukocytes 10.07 K/µL 3.90-10.50 K/µL
Platelets 167 K/µL 150-370 K/µL
MCV 95.4 fL 80.0-99.0 fL
MCH 32.2 pg 27.0-33.5 pg
MCHC 33.7 g/dL 31.5-36.0 g/dL
MPV 11.7 fL 7.0-12.0 fL
RDW-CV 12.6% 11.5-15.0%
Prothrombin Time 120% 78-123%
INR 0.94 0.90-1.25
aPTT 30.1 sec 25.0-38.0 sec
**Addition: Histology Report:**
[Microscopy:]{.underline} (Hematoxylin and Eosin staining):
Histologically, infiltrates of a mesenchymal neoplasm can be seen in a
section we prepared. Below this are areas estimated to contain 90%
tumor, which have been selected and labeled for molecular pathological
analysis.
[Molecular Pathology:]{.underline} After macrodissection of the marked
tumor areas from unstained consecutive sections, RNA was extracted and
analyzed using focused Next-Generation Sequencing technology. The
examination was performed using the FusioPlex Sarcoma v2 Assays, that
allows for the detection of fusions involving 63 genes.
[Diagnosis:]{.underline}
1. Incisional biopsy from a myxoid liposarcoma, Grade 1 according to
FNCLCC (Sum score 2 + 0 + 1 = 3), with the detection of a FUS: DDIT3
fusion transcript (right adductor compartment).
2. Predominantly mature fatty tissue as well as fascial tissue.
- In addition to previous reports, myxoid neoplasm is characterized by
minimal cell density/round cell areas, here less than 25%, according
to FNCLCC (=2 points for tumor differentiation).
- No evidence of necrosis (=0 points).
- 2 mitotic figures in 10 high-power fields (=1 point).
- Total score is 2 + 0 + 1 = 3, corresponding to Grade 1 according to
FNCLCC.
[Diagnosis]{.underline}
1. Incisional biopsy from a myxoid liposarcoma (right adductor
compartment).
2. Predominantly mature fatty tissue as well as fascial tissue
(subcutaneous).
[Comment]{.underline}: The present biopsy material corresponds to Grade
1 according to FNCLCC. A supplementary report follows.
**Supplementary Report from: 03/29/2017:**
[Clinical Information:]{.underline} Suspected liposarcoma of the right
proximal thigh. Encapsulated subfascial tumor, palpably indurated.
Adipose tissue adjacent to the tumor, macroscopically lighter and finer
than the subcutaneous adipose tissue towards the skin.
[Material]{.underline}: Microscopy and Molecular Pathology Interphase
FISH analysis using a two-color break-apart probe to examine a
chromosomal break in the FUS gene (chromosome 16p11.2) and in the DDIT3
gene (chromosome 12q13.3-q14.1).
Interphase FISH analysis reveals a specific break event in the FUS gene
(FUS-FISH positive). This indicates the presence of a FUS translocation.
Similarly, in interphase FISH analysis, a specific break event is
detectable in the DDIT3 gene (DDIT3-FISH positive), indicating the
presence of a DDIT3 translocation.
[Diagnosis:]{.underline} Incisional biopsy from a myxoid liposarcoma of
the right adductor compartment.
Predominantly mature fatty tissue as well as fascial tissue.
[Comment]{.underline}: The cytogenetic findings are indicative of a
myxoid liposarcoma. Technical validation by RNA sequencing will be
provided in a supplemental report. This does not affect the above
diagnosis.
**Supplementary Report from: 03/18/2017:**
[Microscopy: MDM2, S100:]{.underline} Partial weak expression of S100
protein by the lesional cells, occasionally including pre-existing
adipocytes. No abnormal expression of MDM2. No abnormal expression of
MDM2 in mature adipose tissue.
[Diagnosis:]{.underline} Incisional biopsy from a myxoid liposarcoma of
the right adductor compartment.
Predominantly mature fatty tissue as well as fascial tissue.
**Main Report from: 03/18/2017**
[Clinical Information:]{.underline} Suspected liposarcoma of the right
proximal thigh, as per MRI 02/19/2017. Encapsulated subfascial tumor,
palpably indurated located in the right adductor compartment. Adipose
tissue adjacent to the tumor, macroscopically lighter and finer than the
subcutaneous adipose tissue towards the skin.
[Macroscopy:]{.underline}
Tumor: Brown, nodular piece of tissue, 20 x 14 x 10 mm, with smooth and
rough surface. Cut surface shiny and mottled, sometimes gray, sometimes
brown.
Subcutaneous adipose tissue: A piece of adipose tissue, 25 x 20 x 5 mm.
[Microscopy:]{.underline}
Moderately cell dense mesenchymal proliferation with a myxoid matrix.
Predominantly round nuclei, moderately dense nuclear chromatin, slight
pleomorphism. Occasional adipocytic cells with univacuolar cytoplasm.
Partially dense, ribbon-like connective tissue as well as mature
univacuolar adipose tissue.
[Diagnosis:]{.underline}
Incisional biopsy suspected of a myxoid liposarcoma. Predominantly
mature fatty tissue as well as fascial tissue.
### Patient Report 2
**Dear colleague, **
We would like to inform you about our patient Mr. John Havers, born on
05/29/1953, who was admitted to our hospital from 03/29/2017 to
04/05/2017.
**Diagnoses**:
- Myxoid liposarcoma on the right medial thigh, pT2 pNX L0 V0 Pn0 G1
R0, Stage IB
- Incisional biopsy on 03/10/2017
**Other Diagnoses:**
- Arterial hypertension
- Non-insulin-dependent diabetes mellitus
- Coronary artery disease with stent placement
- Nicotine abuse (80-100 pack-years)
- Arterial hypertension
- Status post apoplexy
- Status post cataract surgery
- Status post right hip total hip replacement
- Status post Polypectomy for polyposis coli (minimal dysplasia)
- Status post appendectomy
- M. Meniere
**Allergies**: Hay fever
**Current Presentation**: Neoplasm of uncertain or unknown behavior.
**Treatment**: On 04/01/2017, en bloc tumor excision with removal of the
old biopsy scar, partial resection of the M. gracilis, fibers of the M.
sartorius and M. adductor longus, and ligation of the V. saphena magna
was performed.
**Histology from 04/11/2017**
Clinical Information: Myxoid liposarcoma, localized in the right thigh.
[Macroscopy Tumor, right thigh]{.underline}: A triple surgical resection
was performed, removing skin and subcutaneous tissue and the underlying
soft tissue and muscle. The size of the excised skin spindle was 130 x
45 mm with a resection depth of up to 48 mm. A wound 25 mm long and 6 mm
wide was noted on the skin surface. The muscle attached
laterally/dorsally measured 75 x 25 x 6 mm. Two nodules were noted on
the cut surface. The larger nodule, located in the subcutaneous tissue,
measured 33 mm (proximal/distal) x 36 mm (anterior/dorsal) x 30 mm. Its
distance from the proximal preparation cap was 26 mm, from the distal
preparation cap more than 60 mm, from the ventral soft tissue 20 mm, and
from the dorsal soft tissue 3 mm, with less than 1 mm basal extension.
Superficially, it was surrounded by a delicate capsule. A separate
nodule measuring 20 mm (proximal/distal) x 24 mm (ventral/dorsal) x 20
mm was found immediately ventro-distal to the first nodule. This nodule
was located more than 40 mm from the proximal preparation cap, more than
50 mm from the distal preparation cap, 12 mm ventrally, 11 mm dorsally,
and 5 mm basally. Consequently, the maximum size of the tumor from
proximal to distal was 53 mm. No macroscopic necrotic areas were evident
on the cut surface of the nodule. However, partial necrosis of the
subcutaneous fatty tissue in the vicinity of the described wound was
observed.
[Microscopy HE, PAS:]{.underline} Histomorphologically, there is a
moderately cell-dense proliferation with a significant myxoid matrix in
the area of the two confluent nodules, with a maximum diameter of 53 mm.
There are also areas with relative cell poverty. Within the myxoid
matrix, there are blood vessels with a distinct growth pattern referred
to as the \"chicken wire pattern.\" No clear tumor necroses are evident.
The tumor cell nuclei have a round configuration with moderately dense
chromatin. Apoptotic figures are increased. The number of mitoses is
low.The lesion was completely removed with a minimal margin of 0.5 mm
from the posterior resection edge. In the superficial subcutaneous
tissue, there is a band-like necrosis directly related to superficial
granulation tissue. The included skin spindle shows regular epidermal
covering and a largely unremarkable dermis.
[Diagnosis]{.underline}: Skin/subcutaneous excision with a maximum 53 mm
myxoid liposarcoma that was completely removed (minimum distance to
posterior cutoff plane 0.5 mm).
[Comment]{.underline}: In view of the present morphology and knowledge
of the molecular pathological examination results with proven break
events in the FUS gene and DDIT3 gene as part of interphase FISH
analysis, the diagnosed condition is myxoid liposarcoma.
According to the FNCLCC grading scheme, this corresponds to grade 1:
Histological type: 2 points + mitotic index 1 point + necrosis index 0
points = 3 points.
ICD-O-3 tumor classification: Myxoid liposarcoma TNM (8th edition): pT2
pNX L0 V0 Pn0 G1 R0
**Medical History:** We assume that you are familiar with Mr. Havers's
medical history, and we refer to our previous correspondence.
**Physical Examination:** Patient in good general condition. Oriented in
all aspects. No cyanosis. No edema. Warm and dry skin. Normal nasal and
pharyngeal findings. Pupils round, equal, and react promptly to light
bilaterally. Moist tongue. Pharynx and buccal mucosa unremarkable. No
jugular vein distension. No carotid bruits heard. Palpation of lymph
nodes unremarkable. Palpation of the thyroid gland unremarkable, freely
movable. Lungs: Normal chest shape, moderately mobile, vesicular breath
sounds.
Heart: Regular heart action, normal rate; heart sounds clear, no
pathological sounds. Abdomen: Peristalsis and bowel sounds normal in all
quadrants; soft abdomen, no tenderness, no palpable masses, liver and
spleen not palpable due to limited access, non-tender kidneys.
Normal peripheral pulses; joints freely movable. Strength, motor
function, and sensation are unremarkable.
**Therapy and Progression**: The patient presented to our surgical
clinic because of a mass in the right proximal thigh. The swelling was
first noticed about 3 months ago and has increased significantly in size
since then. MRI findings raised suspicion of a liposarcoma. After
consultation in the interdisciplinary tumor board, the indication for
incisional biopsy was performed on 03/10/2017. The histopathological
examination confirmed the presence of a myxoid liposarcoma, leading to
the decision for en bloc excision. The patient was extensively informed
about the procedure and the risks and gave his consent. The patient was
admitted for the procedure on 03/29/2017.
Upon clinical examination, a patient in good general and nutritional
condition was noted. Other general clinical findings were unremarkable.
A wound healing disorder of 2 cm was observed in the area of the wound
after incisional biopsy.
**Sarcoma Tumor Board Recommendation dated 03/11/2017:** R0 G1 finding,
standard sarcoma follow-up.
**Procedure**: Following standard preoperative preparations and informed
consent, the aforementioned procedure was performed on 03/01/2017 under
general anesthesia. The intraoperative and postoperative course was
uneventful.
On the first postoperative day, there was slight swelling in the
affected area, which gradually subsided. Analgesia was sufficient with
Acetaminophen as needed. Thrombosis prophylaxis was administered with
subcutaneous Enoxaparin 0.4 mL. The patient mobilized independently on
the ward. The inserted drainage could not be removed so far due to
excessive drainage output. During the hospital stay, a staging CT of the
chest and abdomen was performed. No thoracoabdominal metastases were
detected.
**Summary**: With a good subjective well-being and unremarkable wound
conditions, Mr. Havers was discharged on 04/05/2017 for further
outpatient care. Clinical examination reveals slight swelling of the
wound area. The wound is not dehiscent and shows no signs of irritation.
The patient is mobilizing independently.
**CT Chest/Abdomen/Pelvis from 04/01/2017: **
[Clinical Information, Question, Justification]{.underline}: Liposarcoma
of the thigh. Staging.
[Technique]{.underline}: Digital overview radiographs. Following
intravenous contrast agent administration (100 ml Xenetix), CT of the
chest and entire abdomen in the venous contrast phase. Reconstruction of
the primary dataset with a slice thickness of 0.625 mm. Multiplanar
reconstruction. Total DLP: 885 mGy\*cm.
[Findings]{.underline}: There are no prior images available for
comparison.
[Chest]{.underline}: Lungs are evenly ventilated and normally developed
bilaterally. No pneumothorax on either side. Minimal right-sided pleural
effusion. Mild basilar hypoventilation, particularly in the right lower
lobe. Calcified granuloma in the apical right lower lobe. No suspicious
pulmonary nodules.
Heart shows enlargement of the left ventricle and left atrium. Coronary
artery sclerosis. Atherosclerosis of the aortic arch. No pericardial
effusion. Aorta and pulmonary trunk have normal diameters. No central
pulmonary artery embolism. No pathologically enlarged mediastinal or
hilar lymph nodes. Symmetric appearance of the neck soft tissues.
Thyroid gland without focal lesions. Axillary lymph nodes are of normal
size.
[Abdomen]{.underline}: Liver is of normal size and has a smooth contour.
No signs of cholestasis. No portal vein thrombosis. No suspicious
intrahepatic lesions. Gallbladder appears normal. Common bile duct is
not dilated. Spleen is not enlarged. Pancreas shows regular lobulation,
and there is no dilatation of the pancreatic duct. Both kidneys are free
from urinary tract obstruction. No solid intrarenal masses. Few renal
cysts. Adrenal glands appear unremarkable. Urinary bladder shows no
focal wall thickening. Prostate is not enlarged. Advanced
atherosclerosis of the abdominal aorta and pelvic vessels. History of
stenting of the left external iliac artery with no reocclusion.
Mesenteric, para-aortic, and parailiac lymph nodes are not
pathologically enlarged. No free intraperitoneal fluid or air is
detected. Osseous Structures: Degenerative changes in the spine. No
evidence of suspicious osseous destruction suggestive of tumors. Soft
tissue mantle appears unremarkable.
**Assessment**: No thoracoabdominal metastases.
**Current Recommendations**:
- Regular wound inspections and dressing changes.
- Documentation of drainage output and removal if the output is \<20
ml/24 hours, expected removal on 04/23/2017 at our outpatient
clinic.
- Removal of sutures is not required for absorbable sutures.
- According to the tumor board decision dated 04/11/2017, we recommend
regular follow-up according to the schedule.
**Sarcoma Follow-up Schedule Stage I**
- Local Follow-up:
1. MRI right thigh: Years 1-5: every 6 months
2. Years 6-10: every 12 months
- Pulmonary Follow-up:
3. Chest X-ray, CT chest with contrast agent Years 1-5: every 6
months in alternation
4. Years 6-10: every 12 months in alternation
**Medication upon Discharge:**
**Medication** **Dosage** **Frequency**
-------------------------------------- ------------ -------------------
Aspirin 100 mg 1-0-0-0
Atorvastatin (Lipitor) 20 mg 0-0-1-0
Enoxaparin (Lovenox) Variable 0-0-1-0
Empagliflozin (Jardiance) 10 mg 1-0-0-0
Metformin Hydrochloride (Glucophage) 1000 mg 1-0-1-0
Metoprolol Tartrate (Lopressor) 50 mg 0.5-0-0.5-0
Acetaminophen (Tylenol) 500 mg 2-2-2-2 if needed
Pantoprazole (Protonix) 20 mg 1-0-0-0
**Lab results upon Discharge:**
**Parameter** **Results** **Reference Range**
------------------------------------------- ------------- ---------------------
Sodium 137 mEq/L 136-145 mEq/L
Potassium 4.4 mEq/L 3.5-4.5 mEq/L
Creatinine 0.74 mg/dL 0.70-1.20 mg/dL
Blood Urea Nitrogen 33 mg/dL 17-48 mg/dL
C-Reactive Protein 1.7 mg/dL \< 5.0 mg/dL
Thyroid-Stimulating Hormone 3.58 mIU/L 0.27-4.20 mIU/L
Hemoglobin 16.5 g/dL 13.5-17.0 g/dL
Hematocrit 49.3% 39.5-50.5%
Red Blood Cells 5.2 M/µL 4.3-5.8 M/µL
White Blood Cells 9.63 K/µL 3.90-10.50 K/µL
Platelets 301 K/µL 150-370 K/µL
Mean Corpuscular Volume 95.7 fL 80.0-99.0 fL
Mean Corpuscular Hemoglobin 32.0 pg 27.0-33.5 pg
Mean Corpuscular Hemoglobin Concentration 33.5 g/dL 31.5-36.0 g/dL
Mean Platelet Volume 10.4 fL 7.0-12.0 fL
Red Cell Distribution Width 12.1% 11.6-14.4%
Activated Partial Thromboplastin Time 32.4 sec 25.0-38.0 sec
### Patient Report 3
**Dear colleague, **
We are writing to provide an update on our patient Mr. John Havers, born
on 05/29/1953, who presented to our outpatient surgery clinic on
04/23/2017.
**Diagnosis**: Myxoid liposarcoma, right medial thigh, pT2 pNX L0 V0 Pn0
G1 R0, Stage IB
- Following incisional biopsy
- After en bloc tumor excision with removal of the previous biopsy
scar, partial resection of the gracilis, sartorius and adductor
longus muscles and ligation of the great saphenous vein.
**Other Diagnoses:**
- Arterial hypertension
- Non-insulin-dependent diabetes mellitus
- Coronary artery disease with stent placement
- Nicotine abuse (80-100 pack-years)
- Arterial hypertension
- Status post apoplexy
- Status post cataract surgery
- Status post right hip total hip replacement
- Status post Polypectomy for polyposis coli (minimal dysplasia)
- Status post appendectomy
- M. Meniere
**Allergies**: Hay fever
**Medical History:** We kindly assume that you are familiar with the
patient\'s detailed medical history and refer to our previous discharge
letter.
**Current Presentation:** The patient presented today for a follow-up
visit in our clinic. He reported no complaints. The Redon drain has not
produced any secretions in the last 2 days.
Clinical examination revealed uneventful wound conditions with applied
Steri-strips. There is no evidence of infection. The Redon drain
contains serous wound secretions.
Procedure: The Redon drain is being removed today. With nearly fully
healed wound conditions, we recommend initiating scar massage with fatty
topical products in the near future.
**MRI of the Right Thigh on** 04/23/2017**:**
[Clinical Background, Question, Justification:]{.underline} Sarcoma
follow-up for myxoid liposarcoma on the right medial thigh, pT2 pNX L0
V0 Pn0 G1 R0, Stage IB. Recurrence? Regional behavior? Lymph nodes?
[Technique]{.underline}: 3 Tesla MRI of the right thigh, both plain and
after the administration of 8 ml of Gadovist intravenously. Supine
position, surface coil.
Sequences: TIRM coronal and axial, T2-TSE coronal and axial, T1 VIBE
Dixon axial, EPI-DWI with ADC map axial, T1-Starvibe vascular images
plain and post-contrast axial with subtraction images, T1-TSE FS
post-contrast coronal.
[Findings]{.underline}: Minor FLAIR hyperintense streaky signal
alteration in the surgical area, most likely scar-related, with slight
diffusion restriction and streaky contrast enhancement. No evidence of a
recurrent suspicious substrate. No nodular contrast enhancement.
Slightly accentuated inguinal lymph nodes on the right, most likely
reactive. Unremarkable visualization of the remaining soft tissue.
Normal bone marrow signal. Bladder filled. Unremarkable representation
of the imaged pelvic organs.
[Assessment]{.underline}: Following the resection of a myxoid
liposarcoma on the right medial thigh, there is a regular postoperative
finding. No indication of local recurrence.
**Chest X-ray in Two Planes on 04/23/2017: **
[Clinical Background, Question, Justification]{.underline}: Myxoid
liposarcoma of the right thigh, initial diagnosis in 2022. Follow-up.
Metastases?
[Findings]{.underline}: No corresponding prior images for comparison.
The upper mediastinum is centrally located and not widened. Hila are
free. No acute congestion. No confluent pneumonic infiltrate. No
evidence of larger intrapulmonary lesions. A 7 mm spot shadow is noted
right suprahilar, primarily representing a vascular structure. No
effusions. No pneumothorax.
**Current Recommendations:** The patient would like to continue
follow-up care with us, so we scheduled an MRI control appointment to
assess the possibility of local recurrence. On this day, a two-view
chest X-ray is also required.
**We recommend the following follow-up schedule:**
- Local Follow-up:
5. MRI right thigh: Years 1-5: every 6 months
6. Years 6-10: every 12 months
- Pulmonary Follow-up:
7. Chest X-ray, CT chest with contrast agent Years 1-5: every 6
months in alternation
8. Years 6-10: every 12 months in alternation
### Patient Report 4
**Dear colleague, **
We are writing to provide an update on our patient Mr. John Havers, born
on 05/29/1953, who presented for tumor follow-up on 02/10/2018, in our
outpatient surgery clinic for a discussion of findings.
**Diagnosis**: Myxoid liposarcoma on the right medial thigh, pT2 pNX L0
V0 Pn0 G1 R0, Stage IB
- Following incisional biopsy
- After en bloc tumor excision with removal of the previous biopsy
scar, partial resection of the gracilis, sartorius and adductor
longus muscles and ligation of the great saphenous vein.
**Other Diagnoses:**
- Arterial hypertension
- Non-insulin-dependent diabetes mellitus
- Coronary artery disease with stent placement
- Nicotine abuse (80-100 pack-years)
- Arterial hypertension
- Status post apoplexy
- Status post cataract surgery
- Status post right hip total hip replacement
- Status post Polypectomy for polyposis coli (minimal dysplasia)
- Status post appendectomy
- M. Meniere
**Summary**: Clinically, there is a regular postoperative finding on the
right thigh.
The control MRI with contrast of the right thigh on 04/23/2017 revealed
morphologically:
- No evidence of a local-regional recurrence.
- In pulmonary follow-up using conventional chest X-ray on 04/23/2017,
no signs of pulmonary metastasis were detected.
**Current Recommendations:** Sarcoma Follow-up Schedule Stage I
- Local Follow-up:
9. MRI right thigh: Years 1-5: every 6 months
10. Years 6-10: every 12 months
- Pulmonary Follow-up:
11. Chest X-ray, CT chest with contrast agent Years 1-5: every 6
months in alternation
12. Years 6-10: every 12 months in alternation
### Patient Report 5
**Dear colleague, **
We are reporting to you on our patient Mr. John Havers, born on
05/29/1953, who presented himself on **08/01/2018** at our outpatient
surgery clinic for a discussion of findings as part of tumor follow-up.
**Diagnosis**: Myxoid liposarcoma, right medial thigh, pT2 pNX L0 V0 Pn0
G1 R0, Stage IB
- Post-incision biopsy
- After en bloc tumor excision with removal of the previous biopsy
scar, partial resection of the gracilis, sartorius and adductor
longus muscles and ligation of the great saphenous vein.
**Other Diagnoses:**
- Arterial hypertension
- Non-insulin-dependent diabetes mellitus (NIDDM)
- Coronary artery disease (CAD) with stent placement
- Nicotine abuse (80-100 pack-years)
- Arterial hypertension
- Status post apoplexy
- Status post cataract surgery
- Status post right hip total hip replacement (THR)
- Status post Polypectomy for polyposis coli (minimal dysplasia)
- Status post appendectomy
- M. Meniere
**Summary**: Clinically, there is a normal postoperative condition in
the right thigh.
**MRI of the Right Thigh on 08/01/2018:**
[Clinical Background, Question, Justification:]{.underline} Sarcoma
follow-up for myxoid liposarcoma on the right medial thigh. Progress
assessment.
[Method]{.underline}: 1.5 Tesla. Localization sequences. TIRM and T2 TSE
coronal. TIRM, T2 TSE, VIBE DIXON, and RESOLVE-DWI axial. StarVIBE FS
before and after contrast + subtraction. T1 TSE FS coronal after
contrast.
[Findings]{.underline}: Comparison with MRI from 04/23/2017.
Post-resection of a myxoid liposarcoma in the proximal medial right
thigh soft tissue. In the surgical area, there is no evidence of a
suspicious nodular, contrast-affine lesion, and no evidence of
malignancy-suspected diffusion restriction. Slight scar-related changes
in the access path. Otherwise, unremarkable presentation of soft tissues
and included bony structures. No inguinal lymphadenopathy. Assessment:
For myxoid liposarcoma, there has been consistent evidence since
02/2018:
**Chest CT on 08/01/2018**:
[Clinical Background, Question, Justification: Liposarcoma on the thigh.
Staging.]{.underline} After risk history assessment, oral and written
explanation of contrast agent application and examination procedure, as
well as potential risks of the examination (see also informed consent
form). Written patient consent.
[Method]{.underline}: Digital overview radiographs. After intravenous
contrast agent administration (80 ml of Imeron), CT of the chest in
venous contrast phase, reconstruction of the primary dataset with a
slice thickness of 0.625 mm. Total DLP 185 mGy\*cm.
[Findings]{.underline}: For comparison, there is a CT of the
chest/abdomen/pelvis from 04/01/2018. No evidence of suspicious
pulmonary nodules. Several partly calcified micronodules bipulmonary,
especially in the right lower lobe (ex. S303/IMA179). Partial
underventilation bipulmonary. No pleural effusion. No evidence of
pathologically enlarged lymph nodes. Constant calcified right hilar
lymph nodes. Calcifying aortic sclerosis along with coronary sclerosis.
Hepatic steatosis. Individual renal cysts. Slightly shrunken left
adrenal gland. Degenerative changes of the axial skeleton without
evidence of a malignancy-suspected osseous lesion.
[Assessment]{.underline}: No evidence of a new thoracic tumor
manifestation.
**Recommendations:** Sarcoma Follow-up
- Local Follow-up:
13. MRI right thigh: Years 1-5: every 6 months
14. Years 6-10: every 12 months
- Pulmonary Follow-up:
15. Chest X-ray, CT chest with contrast agent Years 1-5: every 6
months in alternation
16. Years 6-10: every 12 months in alternation
**Lab results upon Discharge: **
**Parameter** **Results** **Reference Range**
---------------------------------------------- ------------- ---------------------
Sodium 138 mEq/L 136-145 mEq/L
Potassium 4.9 mEq/L 3.5-4.5 mEq/L
Creatinine 0.81 mg/dL 0.70-1.20 mg/dL
Estimated GFR \- \-
Blood Urea Nitrogen 38 mg/dL 17-48 mg/dL
C-Reactive Protein 2.6 mg/dL \< 5.0 mg/dL
Hemoglobin 16.7 g/dL 13.5-17.0 g/dL
Hematocrit 49.5% 39.5-50.5%
RBC 5.2 M/µL 4.3-5.8 M/µL
WBC 10.07 K/µL 3.90-10.50 K/µL
Platelets 167 K/µL 150-370 K/µL
MCV 95.4 fL 80.0-99.0 fL
MCH 32.2 pg 27.0-33.5 pg
MCHC 33.7 g/dL 31.5-36.0 g/dL
MPV 11.7 fL 7.0-12.0 fL
RDW-CV 12.6% 11.5-15.0%
Prothrombin Time 120% 78-123%
International Normalized Ratio (INR) 0.94 0.90-1.25
Activated Partial Thromboplastin Time (aPTT) 30.1 sec 25.0-38.0 sec
### Patient Report 6
**Dear colleague, **
We are writing to provide an update on our patient Mr. John Havers, born
on 05/29/1953, who was admitted to our clinic from 08/14/2023 to
09/02/2023.
**Diagnosis:** Pulmonary Metastasis from Myxoid Liposarcoma
- Myxoid liposarcoma on the right medial thigh, pT2 pNX L0 V0 Pn0 G1
R0, Stage IB
<!-- -->
- Post-incision biopsy
- After en bloc tumor excision with removal of the previous biopsy
scar, partial resection of the gracilis, sartorius and adductor
longus muscles and ligation of the great saphenous vein.
**Other Diagnoses:**
- Arterial hypertension
- Non-insulin-dependent diabetes mellitus (NIDDM)
- Coronary artery disease (CAD) with stent placement
- Nicotine abuse (80-100 pack-years)
- Arterial hypertension
- Status post apoplexy
- Status post cataract surgery
- Status post right hip total hip replacement (THR)
- Status post Polypectomy for polyposis coli (minimal dysplasia)
- Status post appendectomy
- M. Meniere
**Medical History:** Mr. Havers has been under our care for myxoid
liposarcoma, which was previously excised from his right medial thigh.
He had a stable postoperative course and was scheduled for regular
follow-up to monitor for any potential recurrence or metastasis.
**Current Presentation:** During a follow-up appointment on 08/14/2023,
Mr. Havers complained of mild shortness of breath, occasional coughing,
and intermittent chest discomfort. He reported no significant weight
loss but noted a decrease in his overall energy levels. Physical
examination revealed decreased breath sounds in the right lung base.
**Physical Examination:** Patient in adequate general condition.
Oriented in all aspects. No cyanosis. No edema. Warm and dry skin.
Normal nasal and pharyngeal findings. Pupils round, equal, and react
promptly to light bilaterally. Moist tongue. Pharynx and buccal mucosa
unremarkable. No jugular vein distension. No carotid bruits heard.
Palpation of lymph nodes unremarkable. Palpation of the thyroid gland
unremarkable, freely movable. Lungs: Normal chest shape, moderately
mobile, decreased breath sounds in the right lung base. Heart: Regular
heart action, normal rate; heart sounds clear, no pathological sounds.
Abdomen: Peristalsis and bowel sounds normal in all quadrants; soft
abdomen, markedly obese, no tenderness, no palpable masses, liver and
spleen not palpable due to limited access, non-tender kidneys. Normal
peripheral pulses; joints freely movable. Strength, motor function, and
sensation is unremarkable.
**Chest X-ray (08/14/2023):** A chest X-ray was performed, which
revealed a suspicious opacity in the right lower lung field.
**CT Chest (08/16/2023):** In light of the chest X-ray findings, a
contrast-enhanced CT scan of the chest was conducted to obtain more
detailed information. The CT imaging demonstrated a well-defined,
irregularly shaped lesion in the right lower lobe of the lung, measuring
approximately 2.5 cm in diameter. The lesion exhibited characteristics
highly suggestive of a metastatic deposit. There were no other
significant abnormalities noted in the chest.
**Histology (08/21/2023):** Based on the CT findings, a CT-guided core
needle biopsy of the pulmonary lesion was performed to confirm the
nature of the lesion. Histopathological examination of the biopsy
specimen confirmed the presence of myxoid liposarcoma cells in the
pulmonary lesion. Immunohistochemical staining for MDM2 and CDK4
supported the diagnosis of metastatic myxoid liposarcoma.
**Treatment Discussion:** Given the diagnosis of a pulmonary metastasis
from myxoid liposarcoma, the case was reviewed in the interdisciplinary
tumor board. The consensus decision was to pursue surgical resection of
the pulmonary metastasis, as it remained localized and resectable. The
patient and his family were informed of the treatment options and
associated risks, and they provided informed consent for the procedure.
**Surgery Report (08/29/2023):** Mr. Havers underwent a right lower
lobectomy with lymph node dissection to remove the pulmonary metastasis.
The procedure was performed by our thoracic surgery team and was
completed without any immediate complications. Intraoperative frozen
section analysis confirmed the presence of metastatic myxoid liposarcoma
in the resected lung tissue.
**Postoperative Course:** Mr. Havers postoperative course was
uneventful, and he demonstrated good respiratory recovery. He was
managed with adequate pain control and underwent chest physiotherapy to
prevent postoperative complications. Pathological examination of the
resected lung tissue confirmed the presence of metastatic myxoid
liposarcoma, with clear surgical margins.
**Current Recommendations:**
1. **Follow-up:** A strict follow-up plan should be established for Mr.
Havers to monitor for any potential recurrence or new metastatic
lesions. This should include regular clinical assessments, chest
imaging, and other relevant investigations.
**Medication upon Discharge:**
**Medication ** **Dosage** **Frequency**
--------------------------------- ------------ ---------------
Empagliflozin (Jardiance) 10 mg 1-0-0-0
Metformin (Glucophage) 1000 mg 1-0-1-0
Atorvastatin (Lipitor) 20 mg 0-0-1-0
Metoprolol Tartrate (Lopressor) 50 mg 0.5-0-0.5-0
Aspirin 100 mg 1-0-0-0
Pantoprazole (Protonix) 20 mg 1-0-0-0
**Lab results upon Discharge: **
**Parameter** **Results** **Reference Range**
--------------------- ------------- ---------------------
Sodium 135 mEq/L 136-145 mEq/L
Potassium 4.4 mEq/L 3.5-4.5 mEq/L
Creatinine 0.82 mg/dL 0.70-1.20 mg/dL
Estimated GFR \- \-
Blood Urea Nitrogen 39 mg/dL 17-48 mg/dL
C-Reactive Protein 2.5 mg/dL \< 5.0 mg/dL
Hemoglobin 16.6 g/dL 13.5-17.0 g/dL
Hematocrit 49.4 % 39.5-50.5 %
RBC 5.1 M/µL 4.3-5.8 M/µL
WBC 10.04 K/µL 3.90-10.50 K/µL
Platelets 166 K/µL 150-370 K/µL
MCV 95.2 fL 80.0-99.0 fL
MCH 32.6 pg 27.0-33.5 pg
MCHC 33.2 g/dL 31.5-36.0 g/dL
MPV 11.4 fL 7.0-12.0 fL
RDW-CV 12.5 % 11.5-15.0 %
Prothrombin Time 122 % 78-123 %
INR 0.99 0.90-1.25
aPTT 30.1 sec 25.0-38.0 sec | 2 mitotic figures |
What is Amazon's year-over-year change in revenue from FY2016 to FY2017 (in units of percents and round to one decimal place)? Calculate what was asked by utilizing the line items clearly shown in the statement of income. | Evidence 0:
Table of Contents
AMAZON.COM, INC.
CONSOLIDATED STATEMENTS OF OPERATIONS
(in millions, except per share data)
Year Ended December 31,
2015
2016
2017
Net product sales
$
79,268 $
94,665 $
118,573
Net service sales
27,738
41,322
59,293
Total net sales
107,006
135,987
177,866
Operating expenses:
Cost of sales
71,651
88,265
111,934
Fulfillment
13,410
17,619
25,249
Marketing
5,254
7,233
10,069
Technology and content
12,540
16,085
22,620
General and administrative
1,747
2,432
3,674
Other operating expense, net
171
167
214
Total operating expenses
104,773
131,801
173,760
Operating income
2,233
4,186
4,106
Interest income
50
100
202
Interest expense
(459)
(484)
(848)
Other income (expense), net
(256)
90
346
Total non-operating income (expense)
(665)
(294)
(300)
Income before income taxes
1,568
3,892
3,806
Provision for income taxes
(950)
(1,425)
(769)
Equity-method investment activity, net of tax
(22)
(96)
(4)
Net income
$
596 $
2,371 $
3,033
Basic earnings per share
$
1.28 $
5.01 $
6.32
Diluted earnings per share
$
1.25 $
4.90 $
6.15
Weighted-average shares used in computation of earnings per share:
Basic
467
474
480
Diluted
477
484
493
See accompanying notes to consolidated financial statements.
38 | 30.8% |
What is their baseline model? | ### Introduction
Social media such as Facebook, Twitter, and Short Text Messaging Service (SMS) are popular channels for getting feedback from consumers on products and services. In Pakistan, with the emergence of e-government practices, SMS is being used for getting feedback from the citizens on different public services with the aim to reduce petty corruption and deficient delivery in services. Automatic classification of these SMS into predefined categories can greatly decrease the response time on complaints and consequently improve the public services rendered to the citizens. While Urdu is the national language of Pakistan, English is treated as the official language of the country. This leads to the development of a distinct dialect of communication known as Roman Urdu, which utilizes English alphabets to write Urdu. Hence, the SMS texts contain multilingual text written in the non-native script and informal diction. The utilization of two or more languages simultaneously is known as multilingualism BIBREF0. Consequently, alternation of two languages in a single conversation, a phenomenon known as code-switching, is inevitable for a multilingual speaker BIBREF1. Factors like informal verbiage, improper grammar, variation in spellings, code-switching, and short text length make the problem of automatic bilingual SMS classification highly challenging. In Natural Language Processing (NLP), deep learning has revolutionized the modeling and understanding of human languages. The richness, expressiveness, ambiguities, and complexity of the natural language can be addressed by deep neural networks without the need to produce complex engineered features BIBREF2. Deep learning models have been successfully used in many NLP tasks involving multilingual text. A Convolutional Neural Network (CNN) based model for sentiment classification of a multilingual dataset was proposed in BIBREF3. However, a particular record in the dataset belonged to one language only. In our case, a record can have either one or two languages. There is very little published work on this specific setting. One way to classify bilingual text is to normalize the different variations of a word to a standard spelling before training the model BIBREF4. However, such normalization requires external resources such as lexical database, and Roman Urdu is under-resourced in this context. Another approach for an under-resourced language is to adapt the resources from resource-rich language BIBREF5. However, such an approach is not generalizable in the case of Roman Urdu text as it is an informal language with no proper grammatical rules and dictionary. More recent approach utilizes code-switching annotations to improve the predictive performance of the model, where each word is annotated with its respective language label. Such an approach is not scalable for large data as annotation task becomes tedious. In this paper, we propose a multi-cascaded deep learning network, called as McM for multi-class classification of bilingual short text. Our goal is to achieve this without any prior knowledge of the language, code-switching indication, language translation, normalizing lexical variations, or language transliteration. In multilingual text classification, previous approaches employ a single deep learning architecture, such as CNN or Long Short Term Memory (LSTM) for feature learning and classification. McM, on the other hand, employs three cascades (aka feature learners) to learn rich textual representations from three perspectives. These representations are then forwarded to a small discriminator network for final prediction. We compare the performance of the proposed model with existing CNN-based model for multilingual text classification BIBREF3. We report a series of experiments using 3 kinds of embedding initialization approaches as well as the effect of attention mechanism BIBREF6. The English language is well studied under the umbrella of NLP, hence many resources and datasets for the different problems are available. However, research on English-Roman Urdu bilingual text lags behind because of non-availability of gold standard datasets. Our second contribution is that we present a large scale annotated dataset in Roman Urdu and English language with code-switching, for multi-class classification. The dataset consists of more than $0.3$ million records and has been made available for future research. The rest of the paper is organized as follows. Section SECREF2 defines the dataset acquiring process and provides an explanation of the class labels. In section SECREF3, the architecture of the proposed model, its hyperparameters, and the experimental setup is discussed. We discuss the results in section SECREF4 and finally, concluding remarks are presented in section SECREF5. . ### Dataset Acquisition and Description
The dataset consists of SMS feedbacks of the citizens of Pakistan on different public services availed by them. The objective of collecting these responses is to measure the performance of government departments rendering different public services. Preprocessing of the data is kept minimal. All records having only single word in SMS were removed as cleaning step. To construct the “gold standard", $313,813$ samples are manually annotated into 12 predefined categories by two annotators in supervision of a domain-expert. Involvement of the domain-expert was to ensure the practicality and quality of the “gold standard". Finally, stratified sampling method was opted for splitting the data into train and test partitions with $80-20$ ratio (i.e., $80\%$ records for training and $20\%$ records for testing). This way, training split has $251,050$ records while testing split has $62,763$ records. The rationale behind stratified sampling was to maintain the ratio of every class in both splits. The preprocessed and annotated data along with train and test split is made available . Note that the department names and service availed by the citizens is mapped to an integer identifier for anonymity. Class label ratios, corresponding labels, and it's description are presented in Table TABREF1. ### Proposed Model and Experimentation
The proposed model, named McM, is mainly inspired by the findings by Reimers, N., & Gurevych (2017) , who concluded that deeper model have minimal effect on the predictive performance of the model BIBREF7. McM manifests a wider model, which employ three feature learners (cascades) that are trained for classification independently (in parallel). The input text is first mapped to embedding matrix of size $l \times d$ where $l$ denotes the number of words in the text while $d$ is dimensions of the embedding vector for each of these words. More formally, let $\mathcal {T} \in \lbrace w_1, w_2, ..., w_l\rbrace $ be the input text with $l$ words, embedding matrix is defined by ${X} \in \mathbb {R}^{l \times d}$. This representation is then fed to three feature learners, which are trained with local supervision. The learned features are then forwarded to discriminator network for final prediction as shown in Fig. FIGREF3. Each of these components are discussed in subsequent subsections. ### Proposed Model and Experimentation ::: Stacked-CNN Learner
CNN learner is employed to learn $n$-gram features for identification of relationships between words. A 1-d convolution filter is used with a sliding window (kernel) of size $k$ (number of $n$-grams) in order to extract the features. A filter $W$ is defined as $W \in \mathbb {R}^{k \times d}$ for the convolution function. The word vectors starting from the position $j$ to the position $j + k -1$ are processed by the filter $W$ at a time. The window $h_j$ is expressed as: Where, the $\oplus $ represents the concatenation of word vectors. The number of filters are usually decided empirically. Each filter convolves with one window at a time to generate a feature map $f_j$ for that specific window as: Where, the $\odot $ represents convolution operation, $b$ is a bias term, and $\sigma $ is a nonlinear transformation function ReLU, which is defined as $\sigma (x) = max(x,0)$. The feature maps of each window are concatenated across all filters to get a high level vector representation and fed as input to next CNN layer. Output of second CNN layer is followed by (i) global max-pooling to remove low activation information from feature maps of all filters, and (ii) global average-pooling to get average activation across all the $n$-grams. These two outputs are then concatenated and forwarded to a small feedforward network having two fully-connected layers, followed by a softmax layer for prediction of this particular learner. Dropout and batch-normalization layers are repeatedly used between both fully-connected layers to avoid features co-adaptation BIBREF8, BIBREF9. ### Proposed Model and Experimentation ::: Stacked-LSTM Learner
The traditional methods in deep learning do not account for previous information while processing current input. LSTM, however, is able to memorize past information and correlate it with current information BIBREF10. LSTM structure has memory cells (aka LSTM cells) that store the information selectively. Each word is treated as one time step and is fed to LSTM in a sequential manner. While processing the input at current time step $X_t$, LSTM also takes into account the previous hidden state $h_{t-1}$. The LSTM represents each time step with an input, a memory, and an output gate, denoted as $i_t, f_t$ and $o_t$ respectively. The hidden state $h_t$ of input $X_t$ for each time step $t$ is given by: Where, the $*$ is element-wise multiplication and $\sigma $ is sigmoid activation function. Stacked-LSTM learner is comprised of two LSTM layers. Let ${H_1}$ be a matrix consisting of output vectors $\lbrace h_1, h_2, ..., h_l\rbrace $ that the first LSTM layer produced, denoting output at each time steps. This matrix is fed to second LSTM layer. Similarly, second layer produces another output matrix $H_2$ which is used to apply global max-pooling and global-average pooling. These two outputs are concatenated and forwarded to a two layered feedforward network for intermediate supervision (prediction), identical to previously described stacked-CNN learner. ### Proposed Model and Experimentation ::: LSTM Learner
LSTM learner is employed to learn long-term dependencies of the text as described in BIBREF10. This learner encodes complete input text recursively. It takes one word vector at a time as input and outputs a single vector. The dimensions of the output vector are equal to the number of LSTM units deployed. This encoded text representation is then forwarded to a small feedforward network, identical to aforementioned two learners, for intermediate supervision in order to learn features. This learner differs from stacked-LSTM learner as it learns sentence features, and not average and max features of all time steps (input words). ### Proposed Model and Experimentation ::: Discriminator Network
The objective of discriminator network is to aggregate features learned by each of above described three learners and squash them into a small network for final prediction. The discriminator employs two fully-connected layers with batch-normalization and dropout layer along with ReLU activation function for non-linearity. The softmax activation function with categorical cross-entropy loss is used on the final prediction layer to get probabilities of each class. The class label is assigned based on maximum probability. This is treated as final prediction of the proposed model. The complete architecture, along with dimensions of each output is shown in Fig. FIGREF3. ### Proposed Model and Experimentation ::: Experimental Setup
Pre-trained word embeddings on massive data, such as GloVe BIBREF11, give boost to predictive performance for multi-class classification BIBREF12. However, such embeddings are limited to English language only with no equivalence for Roman Urdu. Therefore, in this study, we avoid using any word-based pre-trained embeddings to give equal treatment to words of each language. We perform three kinds of experiments. (1) Embedding matrix is constructed using ELMo embeddings BIBREF13, which utilizes characters to form word vectors and produces a word vector with $d = 1024$. We call this variation of the model McM$_\textsubscript {E}$. (2) Embedding matrix is initialized randomly for each word with word vector of size $d = 300$. We refer this particular model as McM$_\textsubscript {R}$. (3) We train domain specific embeddings using word2vec with word vector of size $d = 300$ as suggested in original study BIBREF14. We refer to this particular model as McM$_\textsubscript {D}$. Furthermore, we also introduce soft-attention BIBREF6 between two layers of CNN and LSTM (in respective feature learner) to evaluate effect of attention on bilingual text classification. Attention mechanism “highlights" (assigns more weight) a particular word that contributes more towards correct classification. We refer to attention based experiments with subscript $A$ for all three embedding initializations. This way, a total of 6 experiments are performed with different variations of the proposed model. To mitigate effect of random initialization of network weights, we fix the random seed across all experiments. We train each model for 20 epochs and create a checkpoint at epoch with best predictive performance on test split. We re-implement the model proposed in BIBREF3, and use it as a baseline for our problem. The rationale behind choosing this particular model as a baseline is it's proven good predictive performance on multilingual text classification. For McM, the choices of number of convolutional filters, number of hidden units in first dense layer, number of hidden units in second dense layer, and recurrent units for LSTM are made empirically. Rest of the hyperparameters were selected by performing grid search using $20\%$ stratified validation set from training set on McM$_\textsubscript {R}$. Available choices and final selected parameters are mentioned in Table TABREF18. These choices remained same for all experiments and the validation set was merged back into training set. ### Proposed Model and Experimentation ::: Evaluation Metrics
We employed the standard metrics that are widely adapted in the literature for measuring multi-class classification performance. These metrics are accuracy, precision, recall, and F1-score, where latter three can be computed using micro-average or macro-average strategies BIBREF15. In micro-average strategy, each instance holds equal weight and outcomes are aggregated across all classes to compute a particular metric. This essentially means that the outcome would be influenced by the frequent class, if class distribution is skewed. In macro-average however, metrics for each class are calculated separately and then averaged, irrespective of their class label occurrence ratio. This gives each class equal weight instead of each instance, consequently favoring the under-represented classes. In our particular dataset, it is more plausible to favor smaller classes (i.e., other than “Appreciation" and “Satisfied") to detect potential complaints. Therefore, we choose to report macro-average values for precision, recall, and F1-score which are defined by (DISPLAY_FORM20), (DISPLAY_FORM21), and (DISPLAY_FORM22) respectively. ### Results and Discussion
Before evaluating the McM, we first tested the baseline model on our dataset. Table TABREF23 presents results of baseline and all variations of our experiments. We focus our discussion on F1-score as accuracy is often misleading for dataset with unbalanced class distribution. However, for completeness sake, all measures are reported. It is observed from the results that baseline model performs worst among all the experiments. The reason behind this degradation in performance can be traced back to the nature of the texts in the datasets (i.e., datasets used in original paper of baseline model BIBREF3 and in our study). The approach in base model measure the performance of the model on multilingual dataset in which there is no code-switching involved. The complete text belongs to either one language or the other. However, in our case, the SMS text can have code-switching between two language, variation of spelling, or non-standard grammar. Baseline model is simple 1 layered CNN model that is unable to tackle such challenges. On the other hand, McM learns the features from multiple perspectives, hence feature representations are richer, which consequently leads to a superior predictive performance. As every learner in McM is also supervised, all 4 components of the proposed model (i.e., stacked-CNN learner, stacked-LSTM learner, LSTM-learner, and discriminator) can also be compared with each other. In our experiments, the best performing variation of the proposed model is McM$_\textsubscript {D}$. On this particular setting, discriminator is able to achieve an F1-score of $0.69$ with precision and recall values of $0.72$ and $0.68$ respectively. Other components of McM also show the highest stats for all performance measures. However, for McM$_\textsubscript {DA}$, a significant reduction in performance is observed, although, attention-based models have been proven to show improvement in performance BIBREF6. Investigating the reason behind this drop in performance is beyond the scope of this study. The model variations trained on ELMo embedding have second highest performance. Discriminator of McM$_\textsubscript {E}$ achieves an F1-score of $0.66$, beating other learners in this experiment. However, reduction in performance is persistent when attention is used for McM$_\textsubscript {EA}$. Regarding the experiments with random embedding initialization, McM$_\textsubscript {R}$ shows similar performance to McM$_\textsubscript {EA}$, while McM$_\textsubscript {RA}$ performs the worst. It is worth noting that in each experiment, discriminator network stays on top or performs equally as compared to other components in terms of F1-score. This is indication that discriminator network is able to learn richer representations of text as compared to methods where only single feature learner is deployed. Furthermore, the results for testing error for each component (i.e., 3 learners and a discriminator network) for all 4 variations of the proposed model are presented in Fig. FIGREF24. It is evident that the least error across all components is achieved by McM$_\textsubscript {D}$ model. Turning now to individual component performance, in ELMo embeddings based two models, lowest error is achieved by discriminator network, closely followed by stacked LSTM learner and stacked-CNN learner, while LSTM learner has the highest error. As far as model variations with random embeddings initializations are concerned, most interesting results are observed. As shown in subplot (c) and (d) in Fig. FIGREF24, McM$_\textsubscript {R}$ and McM$_\textsubscript {RA}$ tend to overfit. After second epoch, the error rate for all components of these two variations tend to increase drastically. However, it shows minimum error for discriminator in both variations, again proving that the features learned through multiple cascades are more robust and hold greater discriminative power. Note that in all 6 variations of experiments, the error of discriminator network is the lowest as compared to other components of McM. Hence it can be deduced that learning features through multiple perspectives and aggregating them for final prediction is more fruitful as compared to single method of learning. ### Concluding Remarks
In this work, a new large-scale dataset and a novel deep learning architecture for multi-class classification of bilingual (English-Roman Urdu) text with code-switching is presented. The dataset is intended for enhancement of petty corruption detection in public offices and provides grounds for future research in this direction. While deep learning architecture is proposed for multi-class classification of bilingual SMS without utilizing any external resource. Three word embedding initialization techniques and soft-attention mechanism is also investigated. The observations from extensive experimentation led us to conclude that: (1) word embeddings vectors generated through characters tend to favor bilingual text classification as compared to random embedding initialization, (2) the attention mechanism tend to decrease the predictive performance of the model, irrespective of embedding types used, (3) using features learned through single perspective yield poor performance for bilingual text with code-switching, (4) training domain specific embeddings on a large corpus and using them to train the model achieves the highest performance. With regards to future work, we intend to investigate the reason behind degradation of model performance with soft-attention. Table 1. Description of class label along with distribution of each class (in %) in the acquired dataset Fig. 1. Multi-cascaded model (McM) for bilingual short text classification (figure best seen in color) Table 2. Hyperparameter tuning, the selection range, and final choice Table 3. Performance evaluation of variations of the proposed model and baseline. Showing highest scores in boldface. Fig. 2. Test error for all three feature learners and discriminator network over the epochs for all 4 variations of the model, showing lowest error for domain specific embeddings while highest for random embedding initialization. | the model proposed in BIBREF3 |
What experiments do the authors present to validate their system? | ### Introduction
QnAMaker aims to simplify the process of bot creation by extracting Question-Answer (QA) pairs from data given by users into a Knowledge Base (KB) and providing a conversational layer over it. KB here refers to one instance of azure search index, where the extracted QA are stored. Whenever a developer creates a KB using QnAMaker, they automatically get all NLP capabilities required to answer user's queries. There are other systems such as Google's Dialogflow, IBM's Watson Discovery which tries to solve this problem. QnAMaker provides unique features for the ease of development such as the ability to add a persona-based chit-chat layer on top of the bot. Additionally, bot developers get automatic feedback from the system based on end-user traffic and interaction which helps them in enriching the KB; we call this feature active-learning. Our system also allows user to add Multi-Turn structure to KB using hierarchical extraction and contextual ranking. QnAMaker today supports over 35 languages, and is the only system among its competitors to follow a Server-Client architecture; all the KB data rests only in the client's subscription, giving users total control over their data. QnAMaker is part of Microsoft Cognitive Service and currently runs using the Microsoft Azure Stack. ### System description ::: Architecture
As shown in Figure FIGREF4, humans can have two different kinds of roles in the system: Bot-Developers who want to create a bot using the data they have, and End-Users who will chat with the bot(s) created by bot-developers. The components involved in the process are: QnAMaker Portal: This is the Graphical User Interface (GUI) for using QnAMaker. This website is designed to ease the use of management APIs. It also provides a test pane. QnaMaker Management APIs: This is used for the extraction of Question-Answer (QA) pairs from semi-structured content. It then passes these QA pairs to the web app to create the Knowledge Base Index. Azure Search Index: Stores the KB with questions and answers as indexable columns, thus acting as a retrieval layer. QnaMaker WebApp: Acts as a layer between the Bot, Management APIs, and Azure Search Index. WebApp does ranking on top of retrieved results. WebApp also handles feedback management for active learning. Bot: Calls the WebApp with the User's query to get results. ### System description ::: Bot Development Process
Creating a bot is a 3-step process for a bot developer: Create a QnaMaker Resource in Azure: This creates a WebApp with binaries required to run QnAMaker. It also creates an Azure Search Service for populating the index with any given knowledge base, extracted from user data Use Management APIs to Create/Update/Delete your KB: The Create API automatically extracts the QA pairs and sends the Content to WebApp, which indexes it in Azure Search Index. Developers can also add persona-based chat content and synonyms while creating and updating their KBs. Bot Creation: Create a bot using any framework and call the WebApp hosted in Azure to get your queries answered. There are Bot-Framework templates provided for the same. ### System description ::: Extraction
The Extraction component is responsible for understanding a given document and extracting potential QA pairs. These QA pairs are in turn used to create a KB to be consumed later on by the QnAMaker WebApp to answer user queries. First, the basic blocks from given documents such as text, lines are extracted. Then the layout of the document such as columns, tables, lists, paragraphs, etc is extracted. This is done using Recursive X-Y cut BIBREF0. Following Layout Understanding, each element is tagged as headers, footers, table of content, index, watermark, table, image, table caption, image caption, heading, heading level, and answers. Agglomerative clustering BIBREF1 is used to identify heading and hierarchy to form an intent tree. Leaf nodes from the hierarchy are considered as QA pairs. In the end, the intent tree is further augmented with entities using CRF-based sequence labeling. Intents that are repeated in and across documents are further augmented with their parent intent, adding more context to resolve potential ambiguity. ### System description ::: Retrieval And Ranking
QnAMaker uses Azure Search Index as it's retrieval layer, followed by re-ranking on top of retrieved results (Figure FIGREF21). Azure Search is based on inverted indexing and TF-IDF scores. Azure Search provides fuzzy matching based on edit-distance, thus making retrieval robust to spelling mistakes. It also incorporates lemmatization and normalization. These indexes can scale up to millions of documents, lowering the burden on QnAMaker WebApp which gets less than 100 results to re-rank. Different customers may use QnAMaker for different scenarios such as banking task completion, answering FAQs on company policies, or fun and engagement. The number of QAs, length of questions and answers, number of alternate questions per QA can vary significantly across different types of content. Thus, the ranker model needs to use features that are generic enough to be relevant across all use cases. ### System description ::: Retrieval And Ranking ::: Pre-Processing
The pre-processing layer uses components such as Language Detection, Lemmatization, Speller, and Word Breaker to normalize user queries. It also removes junk characters and stop-words from the user's query. ### System description ::: Retrieval And Ranking ::: Features
Going into granular features and the exact empirical formulas used is out of the scope of this paper. The broad level features used while ranking are: WordNet: There are various features generated using WordNet BIBREF2 matching with questions and answers. This takes care of word-level semantics. For instance, if there is information about “price of furniture" in a KB and the end-user asks about “price of table", the user will likely get a relevant answer. The scores of these WordNet features are calculated as a function of: Distance of 2 words in the WordNet graph Distance of Lowest Common Hypernym from the root Knowledge-Base word importance (Local IDFs) Global word importance (Global IDFs) This is the most important feature in our model as it has the highest relative feature gain. CDSSM: Convolutional Deep Structured Semantic Models BIBREF3 are used for sentence-level semantic matching. This is a dual encoder model that converts text strings (sentences, queries, predicates, entity mentions, etc) into their vector representations. These models are trained using millions of Bing Query Title Click-Through data. Using the source-model for vectorizing user query and target-model for vectorizing answer, we compute the cosine similarity between these two vectors, giving the relevance of answer corresponding to the query. TF-IDF: Though sentence-to-vector models are trained on huge datasets, they fail to effectively disambiguate KB specific data. This is where a standard TF-IDF BIBREF4 featurizer with local and global IDFs helps. ### System description ::: Retrieval And Ranking ::: Contextual Features
We extend the features for contextual ranking by modifying the candidate QAs and user query in these ways: $Query_{modified}$ = Query + Previous Answer; For instance, if user query is “yes" and the previous answer is “do you want to know about XYZ", the current query becomes “do you want to know about XYZ yes". Candidate QnA pairs are appended with its parent Questions and Answers; no contextual information is used from the user's query. For instance, if a candidate QnA has a question “benefits" and its parent question was “know about XYZ", the candidate QA's question is changed to “know about XYZ benefits". The features mentioned in Section SECREF20 are calculated for the above combinations also. These features carry contextual information. ### System description ::: Retrieval And Ranking ::: Modeling and Training
We use gradient-boosted decision trees as our ranking model to combine all the features. Early stopping BIBREF5 based on Generality-to-Progress ratio is used to decide the number of step trees and Tolerant Pruning BIBREF6 helps prevent overfitting. We follow incremental training if there is small changes in features or training data so that the score distribution is not changed drastically. ### System description ::: Persona Based Chit-Chat
We add support for bot-developers to directly enable handling chit-chat queries like “hi", “thank you", “what's up" in their QnAMaker bots. In addition to chit-chat, we also give bot developers the flexibility to ground responses for such queries in a specific personality: professional, witty, friendly, caring, or enthusiastic. For example, the “Humorous" personality can be used for a casual bot, whereas a “Professional" personality is more suited in case of banking FAQs or task-completion bots. There is a list of 100+ predefined intents BIBREF7. There is a curated list of queries for each of these intents, along with a separate query understanding layer for ranking these intents. The arbitration between chit-chat answers and user's knowledge base answers is handled by using a chat-domain classifier BIBREF8. ### System description ::: Active Learning
The majority of the KBs are created using existing FAQ pages or manuals but to improve the quality it requires effort from the developers. Active learning generates suggestions based on end-user feedback as well as ranker's implicit signals. For instance, if for a query, CDSSM feature was confident that one QnA should be ranked higher whereas wordnet feature thought other QnA should be ranked higher, active learning system will try to disambiguate it by showing this as a suggestion to the bot developer. To avoid showing similar suggestions to developers, DB-Scan clustering is done which optimizes the number of suggestions shown. ### Evaluation and Insights
QnAMaker is not domain-specific and can be used for any type of data. To support this claim, we measure our system's performance for datasets across various domains. The evaluations are done by managed judges who understands the knowledge base and then judge user queries relevance to the QA pairs (binary labels). Each query-QA pair is judged by two judges. We filter out data for which judges do not agree on the label. Chit-chat in itself can be considered as a domain. Thus, we evaluate performance on given KB both with and without chit-chat data (last two rows in Table TABREF19), as well as performance on just chit-chat data (2nd row in Table TABREF19). Hybrid of deep learning(CDSSM) and machine learning features give our ranking model low computation cost, high explainability and significant F1/AUC score. Based on QnAMaker usage, we observed these trends: Around 27% of the knowledge bases created use pre-built persona-based chitchat, out of which, $\sim $4% of the knowledge bases are created for chit-chat alone. The highest used personality is Professional which is used in 9% knowledge bases. Around $\sim $25% developers have enabled active learning suggestions. The acceptance to reject ratio for active learning suggestions is 0.31. 25.5% of the knowledge bases use one URL as a source while creation. $\sim $41% of the knowledge bases created use different sources like multiple URLs. 15.19% of the knowledge bases use both URL and editorial content as sources. Rest use just editorial content. ### Demonstration
We demonstrate QnAMaker: a service to add a conversational layer over semi-structured user data. In addition to query-answering, we support novel features like personality-grounded chit-chat, active learning based on user-interaction feedback (Figure FIGREF40), and hierarchical extraction for multi-turn conversations (Figure FIGREF41). The goal of the demonstration will be to show how easy it is to create an intelligent bot using QnAMaker. All the demonstrations will be done on the production website Demo Video can be seen here. ### Future Work
The system currently doesn't highlight the answer span and does not generate answers taking the KB as grounding. We will be soon supporting Answer Span BIBREF9 and KB-grounded response generation BIBREF10 in QnAMaker. We are also working on user-defined personas for chit-chat (automatically learned from user-documents). We aim to enhance our extraction to be able to work for any unstructured document as well as images. We are also experimenting on improving our ranking system by using semantic vector-based search as our retrieval and transformer-based models for re-ranking. Figure 1: Interactions between various components of QnaMaker, along with their scopes: server-side and client-side Table 1: Retrieval And Ranking Measurements Figure 2: QnAMaker Runtime Pipeline Figure 3: Active Learning Suggestions Figure 4: Multi-Turn Knowledge Base | we measure our system's performance for datasets across various domains, evaluations are done by managed judges who understands the knowledge base and then judge user queries relevance to the QA pairs |
Who seemed to get the least annoyed at the restaurant?
A. the man who ordered cold cuts
B. the lady in the evening gown
C. the waiter
D. the bartender
| I am a Nucleus By STEPHEN BARR Illustrated by GAUGHAN [Transcriber's Note: This etext was produced from Galaxy Science Fiction February 1957. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed.] No doubt whatever about it, I had the Indian sign on me ... my comfortably untidy world had suddenly turned into a monstrosity of order! When I got home from the office, I was not so much tired as beaten down, but the effect is similar. I let myself into the apartment, which had an absentee-wife look, and took a cold shower. The present downtown temperature, according to the radio, was eighty-seven degrees, but according to my Greenwich Village thermometer, it was ninety-six. I got dressed and went into the living room, and wished ardently that my wife Molly were here to tell me why the whole place looked so woebegone. What do they do, I asked myself, that I have left undone? I've vacuumed the carpet, I've dusted and I've straightened the cushions.... Ah! The ashtrays. I emptied them, washed them and put them back, but still the place looked wife-deserted. It had been a bad day; I had forgotten to wind the alarm clock, so I'd had to hurry to make a story conference at one of the TV studios I write for. I didn't notice the impending rain storm and had no umbrella when I reached the sidewalk, to find myself confronted with an almost tropical downpour. I would have turned back, but a taxi came up and a woman got out, so I dashed through the rain and got in. "Madison and Fifty-fourth," I said. "Right," said the driver, and I heard the starter grind, and then go on grinding. After some futile efforts, he turned to me. "Sorry, Mac. You'll have to find another cab. Good hunting." If possible, it was raining still harder. I opened my newspaper over my hat and ran for the subway: three blocks. Whizzing traffic held me up at each crossing and I was soaked when I reached the platform, just in time to miss the local. After an abnormal delay, I got one which exactly missed the express at Fourteenth Street. The same thing happened at both ends of the crosstown shuttle, but I found the rain had stopped when I got out at Fifty-first and Lexington. As I walked across to Madison Avenue, I passed a big excavation where they were getting ready to put up a new office building. There was the usual crowd of buffs watching the digging machines and, in particular, a man with a pneumatic drill who was breaking up some hard-packed clay. While I looked, a big lump of it fell away, and for an instant I was able to see something that looked like a chunk of dirty glass, the size of an old-fashioned hatbox. It glittered brilliantly in the sunlight, and then his chattering drill hit it. There was a faint bang and the thing disintegrated. It knocked him on his back, but he got right up and I realized he was not hurt. At the moment of the explosion—if so feeble a thing can be called one—I felt something sting my face and, on touching it, found blood on my hand. I mopped at it with my handkerchief but, though slight, the bleeding would not stop, so I went into a drugstore and bought some pink adhesive which I put on the tiny cut. When I got to the studio, I found that I had missed the story conference. During the day, by actual count, I heard the phrase "I'm just spitballing" eight times, and another Madison Avenue favorite, "The whole ball of wax," twelve times. However, my story had been accepted without change because nobody had noticed my absence from the conference room. There you have what is known as the Advertising World, the Advertising game or the advertising racket, depending upon which rung of the ladder you have achieved. The subway gave a repeat performance going home, and as I got to the apartment house we live in, the cop on the afternoon beat was standing there talking to the doorman. He said, "Hello, Mr. Graham. I guess you must have just have missed it at your office building." I looked blank and he explained, "We just heard it a little while ago: all six elevators in your building jammed at the same time. Sounds crazy. I guess you just missed it." Anything can happen in advertising, I thought. "That's right, Danny, I just missed it," I said, and went on in. Psychiatry tells us that some people are accident-prone; I, on the other hand, seemed recently to be coincidence-prone, fluke-happy, and except for the alarm clock, I'd had no control over what had been going on. I went into our little kitchen to make a drink and reread the directions Molly had left, telling me how to get along by myself until she got back from her mother's in Oyster Bay, a matter of ten days. How to make coffee, how to open a can, whom to call if I took sick and such. My wife used to be a trained nurse and she is quite convinced that I cannot take a breath without her. She is right, but not for the reasons she supposes. I opened the refrigerator to get some ice and saw another notice: "When you take out the Milk or Butter, Put it Right Back. And Close the Door, too." Intimidated, I took my drink into the living room and sat down in front of the typewriter. As I stared at the novel that was to liberate me from Madison Avenue, I noticed a mistake and picked up a pencil. When I put it down, it rolled off the desk, and with my eyes on the manuscript, I groped under the chair for it. Then I looked down. The pencil was standing on its end. There, I thought to myself, is that one chance in a million we hear about, and picked up the pencil. I turned back to my novel and drank some of the highball in hopes of inspiration and surcease from the muggy heat, but nothing came. I went back and read the whole chapter to try to get a forward momentum, but came to a dead stop at the last sentence. Damn the heat, damn the pencil, damn Madison Avenue and advertising. My drink was gone and I went back to the kitchen and read Molly's notes again to see if they would be like a letter from her. I noticed one that I had missed, pinned to the door of the dumbwaiter: "Garbage picked up at 6:30 AM so the idea is to Put it Here the Night Before. I love you." What can you do when the girl loves you? I made another drink and went and stared out of the living room window at the roof opposite. The Sun was out again and a man with a stick was exercising his flock of pigeons. They wheeled in a circle, hoping to be allowed to perch, but were not allowed to. Pigeons fly as a rule in formation and turn simultaneously, so that their wings all catch the sunlight at the same time. I was thinking about this decorative fact when I saw that as they were making a turn, they seemed to bunch up together. By some curious chance, they all wanted the same place in the sky to turn in, and several collided and fell. The man was as surprised as I and went to one of the dazed birds and picked it up. He stood there shaking his head from side to side, stroking its feathers. My speculations about this peculiar aerial traffic accident were interrupted by loud voices in the hallway. Since our building is usually very well behaved, I was astonished to hear what sounded like an incipient free-for-all, and among the angry voices I recognized that of my neighbor, Nat, a very quiet guy who works on a newspaper and has never, to my knowledge, given wild parties, particularly in the late afternoon. "You can't say a thing like that to me!" I heard him shout. "I tell you I got that deck this afternoon and they weren't opened till we started to play!" Several other loud voices started at the same time. "Nobody gets five straight-flushes in a row!" "Yeah, and only when you were dealer!" The tone of the argument was beginning to get ugly, and I opened the door to offer Nat help if he needed it. There were four men confronting him, evidently torn between the desire to make an angry exit and the impulse to stay and beat him up. His face was furiously red and he looked stunned. "Here!" he said, holding out a deck of cards, "For Pete's sake, look at 'em yourselves if you think they're marked!" The nearest man struck them up from his hand. "Okay, Houdini! So they're not marked! All I know is five straight...." His voice trailed away. He and the others stared at the scattered cards on the floor. About half were face down, as might be expected, and the rest face up—all red. Someone must have rung, because at that moment the elevator arrived and the four men, with half frightened, incredulous looks, and in silence, got in and were taken down. My friend stood looking at the neatly arranged cards. "Judas!" he said, and started to pick them up. "Will you look at that! My God, what a session...." I helped him and said to come in for a drink and tell me all about it, but I had an idea what I would hear. After a while, he calmed down, but he still seemed dazed. "Never seen anything to equal it," he said. "Wouldn't have believed it. Those guys didn't believe it. Every round normal, nothing unusual about the hands—three of a kind, a low straight, that sort of thing and one guy got queens over tens, until it gets to be my deal. Brother! Straight flush to the king—every time! And each time, somebody else has four aces...." He started to sweat again, so I got up to fix him another drink. There was one quart of club soda left, but when I tried to open it, the top broke and glass chips got into the bottle. "I'll have to go down for more soda," I said. "I'll come, too. I need air." At the delicatessen on the corner, the man gave me three bottles in what must have been a wet bag, because as he handed them to me over the top of the cold-meat display, the bottom gave and they fell onto the tile floor. None of them broke, although the fall must have been from at least five feet. Nat was too wound up in his thoughts to notice and I was getting used to miracles. We left the proprietor with his mouth open and met Danny, the cop, looking in at the door, also with his mouth open. On the sidewalk, a man walking in front of Nat stooped suddenly to tie his shoe and Nat, to avoid bumping him, stepped off the curb and a taxi swerved to avoid Nat. The street was still wet and the taxi skidded, its rear end lightly flipping the front of one of those small foreign cars, which was going rather fast. It turned sideways and, without any side-slip, went right up the stoop of a brownstone opposite, coming to rest with its nose inside the front door, which a man opened at that moment. The sight of this threw another driver into a skid, and when he and the taxi had stopped sliding around, they were face to face, arranged crosswise to the street. This gave them exactly no room to move either forward or backward, for the car had its back to a hydrant and the taxi to a lamp. Although rather narrow, this is a two-way street, and in no time at all, traffic was stacked up from both directions as far as the avenues. Everyone was honking his horn. Danny was furious—more so when he tried to put through a call to his station house from the box opposite. It was out of order. Upstairs, the wind was blowing into the apartment and I closed the windows, mainly to shut out the tumult and the shouting. Nat had brightened up considerably. "I'll stay for one more drink and then I'm due at the office," he said. "You know, I think this would make an item for the paper." He grinned and nodded toward the pandemonium. When he was gone, I noticed it was getting dark and turned on the desk lamp. Then I saw the curtains. They were all tied in knots, except one. That was tied in three knots. All right , I told myself, it was the wind. But I felt the time had come for me to get expert advice, so I went to the phone to call McGill. McGill is an assistant professor of mathematics at a university uptown and lives near us. He is highly imaginative, but we believe he knows everything. When I picked up the receiver, the line sounded dead and I thought, more trouble. Then I heard a man cough and I said hello. McGill's voice said, "Alec? You must have picked up the receiver just as we were connected. That's a damn funny coincidence." "Not in the least," I said. "Come on over here. I've got something for you to work on." "Well, as a matter of fact, I was calling up to ask you and Molly—" "Molly's away for the week. Can you get over here quick? It's urgent." "At once," he said, and hung up. While I waited, I thought I might try getting down a few paragraphs of my novel—perhaps something would come now. It did, but as I came to a point where I was about to put down the word "agurgling," I decided it was too reminiscent of Gilbert and Sullivan, and stopped at the letter "R." Then I saw that I had unaccountably hit all four keys one step to the side of the correct ones, and tore out the page, with my face red. This was absolutely not my day. "Well," McGill said, "nothing you've told me is impossible or supernatural. Just very, very improbable. In fact, the odds against that poker game alone would lead me to suspect Nat, well as I know him. It's all those other things...." He got up and walked over to the window and looked at the hot twilight while I waited. Then he turned around; he had a look of concern. "Alec, you're a reasonable guy, so I don't think you'll take offense at what I'm going to say. What you have told me is so impossibly unlikely, and the odds against it so astronomical, that I must take the view that you're either stringing me or you're subject to a delusion." I started to get up and expostulate, but he motioned me back. "I know, but don't you see that that is far more likely than...." He stopped and shook his head. Then he brightened. "I have an idea. Maybe we can have a demonstration." He thought for a tense minute and snapped his fingers. "Have you any change on you?" "Why, yes," I said. "Quite a bit." I reached into my pocket. There must have been nearly two dollars in silver and pennies. "Do you think they'll each have the same date, perhaps?" "Did you accumulate all that change today?" "No. During the week." He shook his head. "In that case, no. Discounting the fact that you could have prearranged it, if my dim provisional theory is right, that would be actually impossible. It would involve time-reversal. I'll tell you about it later. No, just throw down the change. Let's see if they all come up heads." I moved away from the carpet and tossed the handful of coins onto the floor. They clattered and bounced—and bounced together—and stacked themselves into a neat pile. I looked at McGill. His eyes were narrowed. Without a word, he took a handful of coins from his own pocket and threw them. These coins didn't stack. They just fell into an exactly straight line, the adjacent ones touching. "Well," I said, "what more do you want?" "Great Scott," he said, and sat down. "I suppose you know that there are two great apparently opposite principles governing the Universe—random and design. The sands on the beach are an example of random distribution and life is an example of design. The motions of the particles of a gas are what we call random, but there are so many of them, we treat them statistically and derive the Second Law of Thermodynamics—quite reliable. It isn't theoretically hard-and-fast; it's just a matter of extreme probability. Now life, on the other hand, seems not to depend on probability at all; actually, it goes against it. Or you might say it is certainly not an accidental manifestation." "Do you mean," I asked in some confusion, "that some form of life is controlling the coins and—the other things?" He shook his head. "No. All I mean is that improbable things usually have improbable explanations. When I see a natural law being broken, I don't say to myself, 'Here's a miracle.' I revise my version of the book of rules. Something—I don't know what—is going on, and it seems to involve probability, and it seems to center around you. Were you still in that building when the elevators stuck? Or near it?" "I guess I must have been. It happened just after I left." "Hm. You're the center, all right. But why?" "Center of what?" I asked. "I feel as though I were the center of an electrical storm. Something has it in for me!" McGill grinned. "Don't be superstitious. And especially don't be anthropomorphic." "Well, if it's the opposite of random, it's got to be a form of life." "On what basis? All we know for certain is that random motions are being rearranged. A crystal, for example, is not life, but it's a non-random arrangement of particles.... I wonder." He had a faraway, frowning look. I was beginning to feel hungry and the drinks had worn off. "Let's go out and eat," I said, "There's not a damn thing in the kitchen and I'm not allowed to cook. Only eggs and coffee." We put on our hats and went down to the street. From either end, we could hear wrecking trucks towing away the stalled cars. There were, by this time, a number of harassed cops directing the maneuver and we heard one of them say to Danny, "I don't know what the hell's going on around here. Every goddam car's got something the matter with it. They can't none of them back out for one reason or another. Never seen anything like it." Near us, two pedestrians were doing a curious little two-step as they tried to pass one another; as soon as one of them moved aside to let the other pass, the other would move to the same side. They both had embarrassed grins on their faces, but before long their grins were replaced by looks of suspicion and then determination. "All right, smart guy!" they shouted in unison, and barged ahead, only to collide. They backed off and threw simultaneous punches which met in mid-air. Then began one of the most remarkable bouts ever witnessed—a fight in which fist hit fist but never anything else, until both champions backed away undefeated, muttering identical excuses and threats. Danny appeared at that moment. His face was dripping. "You all right, Mr. Graham?" he asked. "I don't know what's going on around here, but ever since I came on this afternoon, things are going crazy. Bartley!" he shouted—he could succeed as a hog-caller. "Bring those dames over here!" Three women in a confused wrangle, with their half-open umbrellas intertwined, were brought across the street, which meant climbing over fenders. Bartley, a fine young patrolman, seemed self-conscious; the ladies seemed not to be. "All right, now, Mrs. Mac-Philip!" one of them said. "Leave go of my umbrella and we'll say no more about it!" "And so now it's Missus Mac-Philip, is it?" said her adversary. The third, a younger one with her back turned to us, her umbrella also caught in the tangle, pulled at it in a tentative way, at which the other two glared at her. She turned her head away and tried to let go, but the handle was caught in her glove. She looked up and I saw it was Molly. My nurse-wife. "Oh, Alec!" she said, and managed to detach herself. "Are you all right?" Was I all right! "Molly! What are you doing here?" "I was so worried, and when I saw all this, I didn't know what to think." She pointed to the stalled cars. "Are you really all right?" "Of course I'm all right. But why...." "The Oyster Bay operator said someone kept dialing and dialing Mother's number and there wasn't anyone on the line, so then she had it traced and it came from our phone here. I kept calling up, but I only got a busy signal. Oh, dear, are you sure you're all right?" I put my arm around her and glanced at McGill. He had an inward look. Then I caught Danny's eye. It had a thoughtful, almost suspicious cast to it. "Trouble does seem to follow you, Mr. Graham," was all he said. When we got upstairs, I turned to McGill. "Explain to Molly," I said. "And incidentally to me. I'm not properly briefed yet." He did so, and when he got to the summing up, I had the feeling she was a jump ahead of him. "In other words, you think it's something organic?" "Well," McGill said, "I'm trying to think of anything else it might be. I'm not doing so well," he confessed. "But so far as I can see," Molly answered, "it's mere probability, and without any over-all pattern." "Not quite. It has a center. Alec is the center." Molly looked at me with a curious expression for a moment. "Do you feel all right, darling?" she asked me. I nodded brightly. "You'll think this silly of me," she went on to McGill, "but why isn't it something like an overactive poltergeist?" "Pure concept," he said. "No genuine evidence." "Magnetism?" "Absolutely not. For one thing, most of the objects affected weren't magnetic—and don't forget magnetism is a force, not a form of energy, and a great deal of energy has been involved. I admit the energy has mainly been supplied by the things themselves, but in a magnetic field, all you'd get would be stored kinetic energy, such as when a piece of iron moves to a magnet or a line of force. Then it would just stay there, like a rundown clock weight. These things do a lot more than that—they go on moving." "Why did you mention a crystal before? Why not a life-form?" "Only an analogy," said McGill. "A crystal resembles life in that it has a definite shape and exhibits growth, but that's all. I'll agree this—thing—has no discernible shape and motion is involved, but plants don't move and amebas have no shape. Then a crystal feeds, but it does not convert what it feeds on; it merely rearranges it into a non-random pattern. In this case, it's rearranging random motions and it has a nucleus and it seems to be growing—at least in what you might call improbability." Molly frowned. "Then what is it? What's it made of?" "I should say it was made of the motions. There's a similar idea about the atom. Another thing that's like a crystal is that it appears to be forming around a nucleus not of its own material—the way a speck of sand thrown into a supersaturated solution becomes the nucleus of crystallization." "Sounds like the pearl in an oyster," Molly said, and gave me an impertinent look. "Why," I asked McGill, "did you say the coins couldn't have the same date? I mean apart from the off chance I got them that way." "Because I don't think this thing got going before today and everything that's happened can all be described as improbable motions here and now. The dates were already there, and to change them would require retroactive action, reversing time. That's out, in my book. That telephone now—" The doorbell rang. We were not surprised to find it was the telephone repairman. He took the set apart and clucked like a hen. "I guess you dropped it on the floor, mister," he said with strong disapproval. "Certainly not," I said. "Is it broken?" "Not exactly broken , but—" He shook his head and took it apart some more. McGill went over and they discussed the problem in undertones. Finally the man left and Molly called her mother to reassure her. McGill tried to explain to me what had happened with the phone. "You must have joggled something loose. And then you replaced the receiver in such a way that the contact wasn't quite open." "But for Pete's sake, Molly says the calls were going on for a long time! I phoned you only a short time ago and it must have taken her nearly two hours to get here from Oyster Bay." "Then you must have done it twice and the vibrations in the floor—something like that—just happened to cause the right induction impulses. Yes, I know how you feel," he said, seeing my expression. "It's beginning to bear down." Molly was through telephoning and suggested going out for dinner. I was so pleased to see her that I'd forgotten all about being hungry. "I'm in no mood to cook," she said. "Let's get away from all this." McGill raised an eyebrow. "If all this, as you call it, will let us." In the lobby, we ran into Nat, looking smug in a journalistic way. "I've been put on the story—who could be better?—I live here. So far, I don't quite get what's been happening. I've been talking to Danny, but he didn't say much. I got the feeling he thinks you're involved in some mystical, Hibernian way. Hello, McGill, what's with you?" "He's got a theory," said Molly. "Come and eat with us and he'll tell you all about it." Since we decided on an air-conditioned restaurant nearby on Sixth Avenue, we walked. The jam of cars didn't seem to be any less than before and we saw Danny again. He was talking to a police lieutenant, and when he caught sight of us, he said something that made the lieutenant look at us with interest. Particularly at me. "If you want your umbrella, Mrs. Graham," Danny said, "it's at the station house. What there's left of it, that is." Molly thanked him and there was a short pause, during which I felt the speculative regard of the lieutenant. I pulled out a packet of cigarettes, which I had opened, as always, by tearing off the top. I happened to have it upside down and all the cigarettes fell out. Before I could move my foot to obliterate what they had spelled out on the sidewalk, the two cops saw it. The lieutenant gave me a hard look, but said nothing. I quickly kicked the insulting cigarettes into the gutter. When we got to the restaurant, it was crowded but cool—although it didn't stay cool for long. We sat down at a side table near the door and ordered Tom Collinses as we looked at the menu. Sitting at the next table were a fat lady, wearing a very long, brilliant green evening gown, and a dried-up sour-looking man in a tux. When the waiter returned, they preempted him and began ordering dinner fussily: cold cuts for the man, and vichyssoise, lobster salad and strawberry parfait for the fat lady. I tasted my drink. It was most peculiar; salt seemed to have been used instead of sugar. I mentioned this and my companions tried theirs, and made faces. The waiter was concerned and apologetic, and took the drinks back to the bar across the room. The bartender looked over at us and tasted one of the drinks. Then he dumped them in his sink with a puzzled expression and made a new batch. After shaking this up, he set out a row of glasses, put ice in them and began to pour. That is to say he tilted the shaker over the first one, but nothing came out. He bumped it against the side of the bar and tried again. Still nothing. Then he took off the top and pried into it with his pick, his face pink with exasperation. I had the impression that the shaker had frozen solid. Well, ice is a crystal, I thought to myself. The other bartender gave him a fresh shaker, but the same thing happened, and I saw no more because the customers sitting at the bar crowded around in front of him, offering advice. Our waiter came back, baffled, saying he'd have the drinks in a moment, and went to the kitchen. When he returned, he had madame's vichyssoise and some rolls, which he put down, and then went to the bar, where the audience had grown larger. Molly lit a cigarette and said, "I suppose this is all part of it, Alec. Incidentally, it seems to be getting warmer in here." It was, and I had the feeling the place was quieter—a background noise had stopped. It dawned on me that I no longer heard the faint hum of the air-conditioner over the door, and as I started to say so, I made a gesture toward it. My hand collided with Molly's when she tapped her cigarette over the ashtray, and the cigarette landed in the neighboring vichyssoise. "Hey! What's the idea?" snarled the sour-looking man. "I'm terribly sorry," I said. "It was an accident. I—" "Throwing cigarettes at people!" the fat lady said. "I really didn't mean to," I began again, getting up. There must have been a hole in the edge of their tablecloth which one of my cuff buttons caught in, because as I stepped out from between the closely set tables, I pulled everything—tablecloth, silver, water glasses, ashtrays and the vichyssoise-à-la-nicotine—onto the floor. The fat lady surged from the banquette and slapped me meatily. The man licked his thumb and danced as boxers are popularly supposed to do. The owner of the place, a man with thick black eyebrows, hustled toward us with a determined manner. I tried to explain what had happened, but I was outshouted, and the owner frowned darkly. | C. the waiter |
What is the most surprising detail about the Venusian delegate?
A. She is very tall for a female
B. It must be assembled according to instructions
C. He was once an inhabitant of Earth
D. It self-destructs after a certain time period has passed
| The saucer was interesting, but where was the delegate? The DELEGATE FROM VENUS By HENRY SLESAR ILLUSTRATOR NOVICK Everybody was waiting to see what the delegate from Venus looked like. And all they got for their patience was the biggest surprise since David clobbered Goliath. " Let me put it this way," Conners said paternally. "We expect a certain amount of decorum from our Washington news correspondents, and that's all I'm asking for." Jerry Bridges, sitting in the chair opposite his employer's desk, chewed on his knuckles and said nothing. One part of his mind wanted him to play it cagey, to behave the way the newspaper wanted him to behave, to protect the cozy Washington assignment he had waited four years to get. But another part of him, a rebel part, wanted him to stay on the trail of the story he felt sure was about to break. "I didn't mean to make trouble, Mr. Conners," he said casually. "It just seemed strange, all these exchanges of couriers in the past two days. I couldn't help thinking something was up." "Even if that's true, we'll hear about it through the usual channels," Conners frowned. "But getting a senator's secretary drunk to obtain information—well, that's not only indiscreet, Bridges. It's downright dirty." Jerry grinned. "I didn't take that kind of advantage, Mr. Conners. Not that she wasn't a toothsome little dish ..." "Just thank your lucky stars that it didn't go any further. And from now on—" He waggled a finger at him. "Watch your step." Jerry got up and ambled to the door. But he turned before leaving and said: "By the way. What do you think is going on?" "I haven't the faintest idea." "Don't kid me, Mr. Conners. Think it's war?" "That'll be all, Bridges." The reporter closed the door behind him, and then strolled out of the building into the sunlight. He met Ruskin, the fat little AP correspondent, in front of the Pan-American Building on Constitution Avenue. Ruskin was holding the newspaper that contained the gossip-column item which had started the whole affair, and he seemed more interested in the romantic rather than political implications. As he walked beside him, he said: "So what really happened, pal? That Greta babe really let down her hair?" "Where's your decorum?" Jerry growled. Ruskin giggled. "Boy, she's quite a dame, all right. I think they ought to get the Secret Service to guard her. She really fills out a size 10, don't she?" "Ruskin," Jerry said, "you have a low mind. For a week, this town has been acting like the 39 Steps , and all you can think about is dames. What's the matter with you? Where will you be when the big mushroom cloud comes?" "With Greta, I hope," Ruskin sighed. "What a way to get radioactive." They split off a few blocks later, and Jerry walked until he came to the Red Tape Bar & Grill, a favorite hangout of the local journalists. There were three other newsmen at the bar, and they gave him snickering greetings. He took a small table in the rear and ate his meal in sullen silence. It wasn't the newsmen's jibes that bothered him; it was the certainty that something of major importance was happening in the capitol. There had been hourly conferences at the White House, flying visits by State Department officials, mysterious conferences involving members of the Science Commission. So far, the byword had been secrecy. They knew that Senator Spocker, chairman of the Congressional Science Committee, had been involved in every meeting, but Senator Spocker was unavailable. His secretary, however, was a little more obliging ... Jerry looked up from his coffee and blinked when he saw who was coming through the door of the Bar & Grill. So did every other patron, but for different reasons. Greta Johnson had that effect upon men. Even the confining effect of a mannishly-tailored suit didn't hide her outrageously feminine qualities. She walked straight to his table, and he stood up. "They told me you might be here," she said, breathing hard. "I just wanted to thank you for last night." "Look, Greta—" Wham! Her hand, small and delicate, felt like a slab of lead when it slammed into his cheek. She left a bruise five fingers wide, and then turned and stalked out. He ran after her, the restaurant proprietor shouting about the unpaid bill. It took a rapid dog-trot to reach her side. "Greta, listen!" he panted. "You don't understand about last night. It wasn't the way that lousy columnist said—" She stopped in her tracks. "I wouldn't have minded so much if you'd gotten me drunk. But to use me, just to get a story—" "But I'm a reporter , damn it. It's my job. I'd do it again if I thought you knew anything." She was pouting now. "Well, how do you suppose I feel, knowing you're only interested in me because of the Senator? Anyway, I'll probably lose my job, and then you won't have any use for me." "Good-bye, Greta," Jerry said sadly. "What?" "Good-bye. I suppose you won't want to see me any more." "Did I say that?" "It just won't be any use. We'll always have this thing between us." She looked at him for a moment, and then touched his bruised cheek with a tender, motherly gesture. "Your poor face," she murmured, and then sighed. "Oh, well. I guess there's no use fighting it. Maybe if I did tell you what I know, we could act human again." "Greta!" "But if you print one word of it, Jerry Bridges, I'll never speak to you again!" "Honey," Jerry said, taking her arm, "you can trust me like a brother." "That's not the idea," Greta said stiffly. In a secluded booth at the rear of a restaurant unfrequented by newsmen, Greta leaned forward and said: "At first, they thought it was another sputnik." " Who did?" "The State Department, silly. They got reports from the observatories about another sputnik being launched by the Russians. Only the Russians denied it. Then there were joint meetings, and nobody could figure out what the damn thing was." "Wait a minute," Jerry said dizzily. "You mean to tell me there's another of those metal moons up there?" "But it's not a moon. That's the big point. It's a spaceship." "A what ?" "A spaceship," Greta said coolly, sipping lemonade. "They have been in contact with it now for about three days, and they're thinking of calling a plenary session of the UN just to figure out what to do about it. The only hitch is, Russia doesn't want to wait that long, and is asking for a hurry-up summit meeting to make a decision." "A decision about what?" "About the Venusians, of course." "Greta," Jerry said mildly, "I think you're still a little woozy from last night." "Don't be silly. The spaceship's from Venus; they've already established that. And the people on it—I guess they're people—want to know if they can land their delegate." "Their what?" "Their delegate. They came here for some kind of conference, I guess. They know about the UN and everything, and they want to take part. They say that with all the satellites being launched, that our affairs are their affairs, too. It's kind of confusing, but that's what they say." "You mean these Venusians speak English?" "And Russian. And French. And German. And everything I guess. They've been having radio talks with practically every country for the past three days. Like I say, they want to establish diplomatic relations or something. The Senator thinks that if we don't agree, they might do something drastic, like blow us all up. It's kind of scary." She shivered delicately. "You're taking it mighty calm," he said ironically. "Well, how else can I take it? I'm not even supposed to know about it, except that the Senator is so careless about—" She put her fingers to her lips. "Oh, dear, now you'll really think I'm terrible." "Terrible? I think you're wonderful!" "And you promise not to print it?" "Didn't I say I wouldn't?" "Y-e-s. But you know, you're a liar sometimes, Jerry. I've noticed that about you." The press secretary's secretary, a massive woman with gray hair and impervious to charm, guarded the portals of his office with all the indomitable will of the U. S. Marines. But Jerry Bridges tried. "You don't understand, Lana," he said. "I don't want to see Mr. Howells. I just want you to give him something." "My name's not Lana, and I can't deliver any messages." "But this is something he wants to see." He handed her an envelope, stamped URGENT. "Do it for me, Hedy. And I'll buy you the flashiest pair of diamond earrings in Washington." "Well," the woman said, thawing slightly. "I could deliver it with his next batch of mail." "When will that be?" "In an hour. He's in a terribly important meeting right now." "You've got some mail right there. Earrings and a bracelet to match." She looked at him with exasperation, and then gathered up a stack of memorandums and letters, his own envelope atop it. She came out of the press secretary's office two minutes later with Howells himself, and Howells said: "You there, Bridges. Come in here." "Yes, sir !" Jerry said, breezing by the waiting reporters with a grin of triumph. There were six men in the room, three in military uniform. Howells poked the envelope towards Jerry, and snapped: "This note of yours. Just what do you think it means?" "You know better than I do, Mr. Howells. I'm just doing my job; I think the public has a right to know about this spaceship that's flying around—" His words brought an exclamation from the others. Howells sighed, and said: "Mr. Bridges, you don't make it easy for us. It's our opinion that secrecy is essential, that leakage of the story might cause panic. Since you're the only unauthorized person who knows of it, we have two choices. One of them is to lock you up." Jerry swallowed hard. "The other is perhaps more practical," Howells said. "You'll be taken into our confidence, and allowed to accompany those officials who will be admitted to the landing site. But you will not be allowed to relay the story to the press until such a time as all correspondents are informed. That won't give you a 'scoop' if that's what you call it, but you'll be an eyewitness. That should be worth something." "It's worth a lot," Jerry said eagerly. "Thanks, Mr. Howells." "Don't thank me, I'm not doing you any personal favor. Now about the landing tonight—" "You mean the spaceship's coming down?" "Yes. A special foreign ministers conference was held this morning, and a decision was reached to accept the delegate. Landing instructions are being given at Los Alamos, and the ship will presumably land around midnight tonight. There will be a jet leaving Washington Airport at nine, and you'll be on it. Meanwhile, consider yourself in custody." The USAF jet transport wasn't the only secrecy-shrouded aircraft that took off that evening from Washington Airport. But Jerry Bridges, sitting in the rear seat flanked by two Sphinx-like Secret Service men, knew that he was the only passenger with non-official status aboard. It was only a few minutes past ten when they arrived at the air base at Los Alamos. The desert sky was cloudy and starless, and powerful searchlights probed the thick cumulus. There were sleek, purring black autos waiting to rush the air passengers to some unnamed destination. They drove for twenty minutes across a flat ribbon of desert road, until Jerry sighted what appeared to be a circle of newly-erected lights in the middle of nowhere. On the perimeter, official vehicles were parked in orderly rows, and four USAF trailer trucks were in evidence, their radarscopes turning slowly. There was activity everywhere, but it was well-ordered and unhurried. They had done a good job of keeping the excitement contained. He was allowed to leave the car and stroll unescorted. He tried to talk to some of the scurrying officials, but to no avail. Finally, he contented himself by sitting on the sand, his back against the grill of a staff car, smoking one cigarette after another. As the minutes ticked off, the activity became more frenetic around him. Then the pace slowed, and he knew the appointed moment was approaching. Stillness returned to the desert, and tension was a tangible substance in the night air. The radarscopes spun slowly. The searchlights converged in an intricate pattern. Then the clouds seemed to part! "Here she comes!" a voice shouted. And in a moment, the calm was shattered. At first, he saw nothing. A faint roar was started in the heavens, and it became a growl that increased in volume until even the shouting voices could no longer be heard. Then the crisscrossing lights struck metal, glancing off the gleaming body of a descending object. Larger and larger the object grew, until it assumed the definable shape of a squat silver funnel, falling in a perfect straight line towards the center of the light-ringed area. When it hit, a dust cloud obscured it from sight. A loudspeaker blared out an unintelligible order, but its message was clear. No one moved from their position. Finally, a three-man team, asbestos-clad, lead-shielded, stepped out from the ring of spectators. They carried geiger counters on long poles before them. Jerry held his breath as they approached the object; only when they were yards away did he appreciate its size. It wasn't large; not more than fifteen feet in total circumference. One of the three men waved a gloved hand. "It's okay," a voice breathed behind him. "No radiation ..." Slowly, the ring of spectators closed tighter. They were twenty yards from the ship when the voice spoke to them. "Greetings from Venus," it said, and then repeated the phrase in six languages. "The ship you see is a Venusian Class 7 interplanetary rocket, built for one-passenger. It is clear of all radiation, and is perfectly safe to approach. There is a hatch which may be opened by an automatic lever in the side. Please open this hatch and remove the passenger." An Air Force General whom Jerry couldn't identify stepped forward. He circled the ship warily, and then said something to the others. They came closer, and he touched a small lever on the silvery surface of the funnel. A door slid open. "It's a box!" someone said. "A crate—" "Colligan! Moore! Schaffer! Lend a hand here—" A trio came forward and hoisted the crate out of the ship. Then the voice spoke again; Jerry deduced that it must have been activated by the decreased load of the ship. "Please open the crate. You will find our delegate within. We trust you will treat him with the courtesy of an official emissary." They set to work on the crate, its gray plastic material giving in readily to the application of their tools. But when it was opened, they stood aside in amazement and consternation. There were a variety of metal pieces packed within, protected by a filmy packing material. "Wait a minute," the general said. "Here's a book—" He picked up a gray-bound volume, and opened its cover. "'Instructions for assembling Delegate,'" he read aloud. "'First, remove all parts and arrange them in the following order. A-1, central nervous system housing. A-2 ...'" He looked up. "It's an instruction book," he whispered. "We're supposed to build the damn thing." The Delegate, a handsomely constructed robot almost eight feet tall, was pieced together some three hours later, by a team of scientists and engineers who seemed to find the Venusian instructions as elementary as a blueprint in an Erector set. But simple as the job was, they were obviously impressed by the mechanism they had assembled. It stood impassive until they obeyed the final instruction. "Press Button K ..." They found button K, and pressed it. The robot bowed. "Thank you, gentlemen," it said, in sweet, unmetallic accents. "Now if you will please escort me to the meeting place ..." It wasn't until three days after the landing that Jerry Bridges saw the Delegate again. Along with a dozen assorted government officials, Army officers, and scientists, he was quartered in a quonset hut in Fort Dix, New Jersey. Then, after seventy-two frustrating hours, he was escorted by Marine guard into New York City. No one told him his destination, and it wasn't until he saw the bright strips of light across the face of the United Nations building that he knew where the meeting was to be held. But his greatest surprise was yet to come. The vast auditorium which housed the general assembly was filled to its capacity, but there were new faces behind the plaques which designated the member nations. He couldn't believe his eyes at first, but as the meeting got under way, he knew that it was true. The highest echelons of the world's governments were represented, even—Jerry gulped at the realization—Nikita Khrushchev himself. It was a summit meeting such as he had never dreamed possible, a summit meeting without benefit of long foreign minister's debate. And the cause of it all, a placid, highly-polished metal robot, was seated blithely at a desk which bore the designation: VENUS. The robot delegate stood up. "Gentlemen," it said into the microphone, and the great men at the council tables strained to hear the translator's version through their headphones, "Gentlemen, I thank you for your prompt attention. I come as a Delegate from a great neighbor planet, in the interests of peace and progress for all the solar system. I come in the belief that peace is the responsibility of individuals, of nations, and now of worlds, and that each is dependent upon the other. I speak to you now through the electronic instrumentation which has been created for me, and I come to offer your planet not merely a threat, a promise, or an easy solution—but a challenge." The council room stirred. "Your earth satellites have been viewed with interest by the astronomers of our world, and we foresee the day when contact between our planets will be commonplace. As for ourselves, we have hitherto had little desire to explore beyond our realm, being far too occupied with internal matters. But our isolation cannot last in the face of your progress, so we believe that we must take part in your affairs. "Here, then, is our challenge. Continue your struggle of ideas, compete with each other for the minds of men, fight your bloodless battles, if you know no other means to attain progress. But do all this without unleashing the terrible forces of power now at your command. Once unleashed, these forces may or may not destroy all that you have gained. But we, the scientists of Venus, promise you this—that on the very day your conflict deteriorates into heedless violence, we will not stand by and let the ugly contagion spread. On that day, we of Venus will act swiftly, mercilessly, and relentlessly—to destroy your world completely." Again, the meeting room exploded in a babble of languages. "The vessel which brought me here came as a messenger of peace. But envision it, men of Earth, as a messenger of war. Unstoppable, inexorable, it may return, bearing a different Delegate from Venus—a Delegate of Death, who speaks not in words, but in the explosion of atoms. Think of thousands of such Delegates, fired from a vantage point far beyond the reach of your retaliation. This is the promise and the challenge that will hang in your night sky from this moment forward. Look at the planet Venus, men of Earth, and see a Goddess of Vengeance, poised to wreak its wrath upon those who betray the peace." The Delegate sat down. Four days later, a mysterious explosion rocked the quiet sands of Los Alamos, and the Venus spacecraft was no more. Two hours after that, the robot delegate, its message delivered, its mission fulfilled, requested to be locked inside a bombproof chamber. When the door was opened, the Delegate was an exploded ruin. The news flashed with lightning speed over the world, and Jerry Bridges' eyewitness accounts of the incredible event was syndicated throughout the nation. But his sudden celebrity left him vaguely unsatisfied. He tried to explain his feeling to Greta on his first night back in Washington. They were in his apartment, and it was the first time Greta had consented to pay him the visit. "Well, what's bothering you?" Greta pouted. "You've had the biggest story of the year under your byline. I should think you'd be tickled pink." "It's not that," Jerry said moodily. "But ever since I heard the Delegate speak, something's been nagging me." "But don't you think he's done good? Don't you think they'll be impressed by what he said?" "I'm not worried about that. I think that damn robot did more for peace than anything that's ever come along in this cockeyed world. But still ..." Greta snuggled up to him on the sofa. "You worry too much. Don't you ever think of anything else? You should learn to relax. It can be fun." She started to prove it to him, and Jerry responded the way a normal, healthy male usually does. But in the middle of an embrace, he cried out: "Wait a minute!" "What's the matter?" "I just thought of something! Now where the hell did I put my old notebooks?" He got up from the sofa and went scurrying to a closet. From a debris of cardboard boxes, he found a worn old leather brief case, and cackled with delight when he found the yellowed notebooks inside. "What are they?" Greta said. "My old school notebooks. Greta, you'll have to excuse me. But there's something I've got to do, right away!" "That's all right with me," Greta said haughtily. "I know when I'm not wanted." She took her hat and coat from the hall closet, gave him one last chance to change his mind, and then left. Five minutes later, Jerry Bridges was calling the airlines. It had been eleven years since Jerry had walked across the campus of Clifton University, heading for the ivy-choked main building. It was remarkable how little had changed, but the students seemed incredibly young. He was winded by the time he asked the pretty girl at the desk where Professor Martin Coltz could be located. "Professor Coltz?" She stuck a pencil to her mouth. "Well, I guess he'd be in the Holland Laboratory about now." "Holland Laboratory? What's that?" "Oh, I guess that was after your time, wasn't it?" Jerry felt decrepit, but managed to say: "It must be something new since I was here. Where is this place?" He followed her directions, and located a fresh-painted building three hundred yards from the men's dorm. He met a student at the door, who told him that Professor Coltz would be found in the physics department. The room was empty when Jerry entered, except for the single stooped figure vigorously erasing a blackboard. He turned when the door opened. If the students looked younger, Professor Coltz was far older than Jerry remembered. He was a tall man, with an unruly confusion of straight gray hair. He blinked when Jerry said: "Hello, Professor. Do you remember me? Jerry Bridges?" "Of course! I thought of you only yesterday, when I saw your name in the papers—" They sat at facing student desks, and chatted about old times. But Jerry was impatient to get to the point of his visit, and he blurted out: "Professor Coltz, something's been bothering me. It bothered me from the moment I heard the Delegate speak. I didn't know what it was until last night, when I dug out my old college notebooks. Thank God I kept them." Coltz's eyes were suddenly hooded. "What do you mean, Jerry?" "There was something about the Robot's speech that sounded familiar—I could have sworn I'd heard some of the words before. I couldn't prove anything until I checked my old notes, and here's what I found." He dug into his coat pocket and produced a sheet of paper. He unfolded it and read aloud. "'It's my belief that peace is the responsibility of individuals, of nations, and someday, even of worlds ...' Sound familiar, Professor?" Coltz shifted uncomfortably. "I don't recall every silly thing I said, Jerry." "But it's an interesting coincidence, isn't it, Professor? These very words were spoken by the Delegate from Venus." "A coincidence—" "Is it? But I also remember your interest in robotics. I'll never forget that mechanical homing pigeon you constructed. And you've probably learned much more these past eleven years." "What are you driving at, Jerry?" "Just this, Professor. I had a little daydream, recently, and I want you to hear it. I dreamed about a group of teachers, scientists, and engineers, a group who were suddenly struck by an exciting, incredible idea. A group that worked in the quiet and secrecy of a University on a fantastic scheme to force the idea of peace into the minds of the world's big shots. Does my dream interest you, Professor?" "Go on." "Well, I dreamt that this group would secretly launch an earth satellite of their own, and arrange for the nose cone to come down safely at a certain time and place. They would install a marvelous electronic robot within the cone, ready to be assembled. They would beam a radio message to earth from the cone, seemingly as if it originated from their 'spaceship.' Then, when the Robot was assembled, they would speak through it to demand peace for all mankind ..." "Jerry, if you do this—" "You don't have to say it, Professor, I know what you're thinking. I'm a reporter, and my business is to tell the world everything I know. But if I did it, there might not be a world for me to write about, would there? No, thanks, Professor. As far as I'm concerned, what I told you was nothing more than a daydream." Jerry braked the convertible to a halt, and put his arm around Greta's shoulder. She looked up at the star-filled night, and sighed romantically. Jerry pointed. "That one." Greta shivered closer to him. "And to think what that terrible planet can do to us!" "Oh, I dunno. Venus is also the Goddess of Love." He swung his other arm around her, and Venus winked approvingly. THE END Transcriber's Note: This etext was produced from Amazing Science Fiction Stories October 1958. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed. Minor spelling and typographical errors have been corrected without note. | B. It must be assembled according to instructions |
How are seed dictionaries obtained by fully unsupervised methods? | ### Introduction and Motivation
The wide use and success of monolingual word embeddings in NLP tasks BIBREF0 , BIBREF1 has inspired further research focus on the induction of cross-lingual word embeddings (CLWEs). CLWE methods learn a shared cross-lingual word vector space where words with similar meanings obtain similar vectors regardless of their actual language. CLWEs benefit cross-lingual NLP, enabling multilingual modeling of meaning and supporting cross-lingual transfer for downstream tasks and resource-lean languages. CLWEs provide invaluable cross-lingual knowledge for, inter alia, bilingual lexicon induction BIBREF2 , BIBREF3 , information retrieval BIBREF4 , BIBREF5 , machine translation BIBREF6 , BIBREF7 , document classification BIBREF8 , cross-lingual plagiarism detection BIBREF9 , domain adaptation BIBREF10 , cross-lingual POS tagging BIBREF11 , BIBREF12 , and cross-lingual dependency parsing BIBREF13 , BIBREF14 . The landscape of CLWE methods has recently been dominated by the so-called projection-based methods BIBREF15 , BIBREF16 , BIBREF17 . They align two monolingual embedding spaces by learning a projection/mapping based on a training dictionary of translation pairs. Besides their simple conceptual design and competitive performance, their popularity originates from the fact that they rely on rather weak cross-lingual supervision. Originally, the seed dictionaries typically spanned several thousand word pairs BIBREF15 , BIBREF18 , BIBREF19 , but more recent work has shown that CLWEs can be induced with even weaker supervision from small dictionaries spanning several hundred pairs BIBREF20 , identical strings BIBREF21 , or even only shared numerals BIBREF22 . Taking the idea of reducing cross-lingual supervision to the extreme, the latest CLWE developments almost exclusively focus on fully unsupervised approaches BIBREF23 , BIBREF24 , BIBREF25 , BIBREF26 , BIBREF27 , BIBREF28 , BIBREF29 , BIBREF30 : they fully abandon any source of (even weak) supervision and extract the initial seed dictionary by exploiting topological similarities between pre-trained monolingual embedding spaces. Their modus operandi can roughly be described by three main components: C1) unsupervised extraction of a seed dictionary; C2) a self-learning procedure that iteratively refines the dictionary to learn projections of increasingly higher quality; and C3) a set of preprocessing and postprocessing steps (e.g., unit length normalization, mean centering, (de)whitening) BIBREF31 that make the entire learning process more robust. The induction of fully unsupervised CLWEs is an inherently interesting research topic per se. Nonetheless, the main practical motivation for developing such approaches in the first place is to facilitate the construction of multilingual NLP tools and widen the access to language technology for resource-poor languages and language pairs. However, the first attempts at fully unsupervised CLWE induction failed exactly for these use cases, as shown by sogaard2018on. Therefore, the follow-up work aimed to improve the robustness of unsupervised CLWE induction by introducing more robust self-learning procedures BIBREF24 , BIBREF32 . Besides increased robustness, recent work claims that fully unsupervised projection-based CLWEs can even match or surpass their supervised counterparts BIBREF23 , BIBREF24 , BIBREF27 , BIBREF33 , BIBREF34 . In this paper, we critically examine these claims on robustness and improved performance of unsupervised CLWEs by running a large-scale evaluation in the bilingual lexicon induction (BLI) task on 15 languages (i.e., 210 languages pairs, see Table 2 in § "Experimental Setup" ). The languages were selected to represent different language families and morphological types, as we argue that fully unsupervised CLWEs have been designed to support exactly these setups. However, we show that even the most robust unsupervised CLWE method BIBREF24 still fails for a large number of language pairs: 87/210 BLI setups are unsuccessful, yielding (near-)zero BLI performance. Further, even when the unsupervised method succeeds, it is because the components C2 (self-learning) and C3 (pre-/post-processing) can mitigate the undesired effects of noisy seed lexicon extraction. We then demonstrate that the combination of C2 and C3 with a small provided seed dictionary (e.g., 500 or 1K pairs) outscores the unsupervised method in all cases, often with a huge margin, and does not fail for any language pair. Furthermore, we show that the most robust unsupervised CLWE approach still fails completely when it relies on monolingual word vectors trained on domain-dissimilar corpora. We also empirically verify that unsupervised approaches cannot outperform weakly supervised approaches also for closely related languages (e.g., Swedish–Danish, Spanish–Catalan). While the “no supervision at all” premise behind fully unsupervised CLWE methods is indeed seductive, our study strongly suggests that future research efforts should revisit the main motivation behind these methods and focus on designing even more robust solutions, given their current inability to support a wide spectrum of language pairs. In hope of boosting induction of CLWEs for more diverse and distant language pairs, we make all 210 training and test dictionaries used in this work publicly available at: https://github.com/ivulic/panlex-bli. ### Methodology
We now dissect a general framework for unsupervised CLWE learning, and show that the “bag of tricks of the trade” used to increase their robustness (which often slips under the radar) can be equally applied to (weakly) supervised projection-based approaches, leading to their fair(er) comparison. ### Projection-Based CLWE Approaches
In short, projection-based CLWE methods learn to (linearly) align independently trained monolingual spaces $\mathbf {X}$ and $\mathbf {Z}$ , using a word translation dictionary $D_0$ to guide the alignment process. Let $\mathbf {X}_D \subset \mathbf {X}$ and $\mathbf {Z}_D \subset \mathbf {Z}$ be the row-aligned subsets of monolingual spaces containing vectors of aligned words from $D_0$ . Alignment matrices $\mathbf {X}_D$ and $\mathbf {Z}_D$ are then used to learn orthogonal transformations $\mathbf {W}_x$ and $\mathbf {W}_z$ that define the joint bilingual space $\mathbf {Z}$0 . While supervised projection-based CLWE models learn the mapping using a provided external (clean) dictionary $\mathbf {Z}$1 , their unsupervised counterparts automatically induce the seed dictionary in an unsupervised way (C1) and then refine it in an iterative fashion (C2). Unsupervised CLWEs. These methods first induce a seed dictionary $D^{(1)}$ leveraging only two unaligned monolingual spaces (C1). While the algorithms for unsupervised seed dictionary induction differ, they all strongly rely on the assumption of similar topological structure between the two pretrained monolingual spaces. Once the seed dictionary is obtained, the two-step iterative self-learning procedure (C2) takes place: 1) a dictionary $D^{(k)}$ is first used to learn the joint space $\mathbf {Y}^{(k)} = \mathbf {X{W}}^{(k)}_x \cup \mathbf {Z{W}}^{(k)}_z$ ; 2) the nearest neighbours in $\mathbf {Y}^{(k)}$ then form the new dictionary $D^{(k+1)}$ . We illustrate the general structure in Figure 1 . A recent empirical survey paper BIBREF17 has compared a variety of latest unsupervised CLWE methods BIBREF23 , BIBREF27 , BIBREF33 , BIBREF24 in several downstream tasks (e.g., BLI, cross-lingual information retrieval, document classification). The results of their study indicate that the vecmap model of artetxe2018robust is by far the most robust and best performing unsupervised CLWE model. For the actual results and analyses, we refer the interested reader to the original paper of glavas2019howto. Another recent evaluation paper BIBREF35 as well as our own preliminary BLI tests (not shown for brevity) have further verified their findings. We thus focus on vecmap in our analyses, and base the following description of the components C1-C3 on that model. ### Three Key Components
C1. Seed Lexicon Extraction. vecmap induces the initial seed dictionary using the following heuristic: monolingual similarity distributions for words with similar meaning will be similar across languages. The monolingual similarity distributions for the two languages are given as rows (or columns; the matrices are symmetric) of $\mathbf {M}_x = \mathbf {X}\mathbf {X}^T$ and $\mathbf {M}_z = \mathbf {Z}\mathbf {Z}^T$ . For the distributions of similarity scores to be comparable, the values in each row of $\mathbf {M}_x$ and $\mathbf {M}_z$ are first sorted. The initial dictionary $D^{(1)}$ is finally obtained by searching for mutual nearest neighbours between the rows of $\sqrt{\mathbf {M}_x}$ and of $\sqrt{\mathbf {M}_z}$ . C2. Self-Learning. Not counting the preprocessing and postprocessing steps (component C3), self-learning then iteratively repeats two steps: 1) Let $\mathbf {D}^{(k)}$ be the binary matrix indicating the aligned words in the dictionary $D^{(k)}$ . The orthogonal transformation matrices are then obtained as $\mathbf {W}^{(k)}_x = \mathbf {U}$ and $\mathbf {W}^{(k)}_z = \mathbf {V}$ , where $\mathbf {U}\mathbf {\Sigma }\mathbf {V}^T$ is the singular value decomposition of the matrix $\mathbf {X}^T\mathbf {D}^{(k)}\mathbf {Z}$ . The cross-lingual space of the $D^{(k)}$0 -th iteration is then $D^{(k)}$1 . 2) The new dictionary $D^{(k+1)}$ is then built by identifying nearest neighbours in $\mathbf {Y}^{(k)}$ . These can be easily extracted from the matrix $\mathbf {P} = \mathbf {X}\mathbf {W}^{(k)}_x( \mathbf {Z}\mathbf {W}^{(k)}_z)^T$ . All nearest neighbours can be used, or additional symmetry constraints can be imposed to extract only mutual nearest neighbours: all pairs of indices ( $i, j$ ) for which $\mathbf {P}_{ij}$ is the largest value both in row $i$ and column $j$ . The above procedure, however, often converges to poor local optima. To remedy for this, the second step (i.e., dictionary induction) is extended with techniques that make self-learning more robust. First, the vocabularies of $\mathbf {X}$ and $\mathbf {Z}$ are cut to the top $k$ most frequent words. Second, similarity scores in $\mathbf {P}$ are kept with probability $p$ , and set to zero otherwise. This dropout allows for a wider exploration of possible word pairs in the dictionary and contributes to escaping poor local optima given the noisy seed lexicon in the first iterations. C3. Preprocessing and Postprocessing Steps. While iteratively learning orthogonal transformations $\mathbf {W}_{x}$ and $\mathbf {W}_{z}$ for $\mathbf {X}$ and $\mathbf {Z}$ is the central step of unsupervised projection-based CLWE methods, preprocessing and postprocessing techniques are additionally applied before and after the transformation. While such techniques are often overlooked in model comparisons, they may have a great impact on the model's final performance, as we validate in § "Results and Discussion" . We briefly summarize two pre-processing (S1 and S2) and post-processing (S3 and S4) steps used in our evaluation, originating from the framework of artetxe2018generalizing. S1) Normalization and mean centering. We first apply unit length normalization: all vectors in $\mathbf {X}$ and $\mathbf {Z}$ are normalized to have a unit Euclidean norm. Following that, $\mathbf {X}$ and $\mathbf {Z}$ are mean centered dimension-wise and then again length-normalized. S2) Whitening. ZCA whitening BIBREF36 is applied on (S1-processed) $\mathbf {X}$ and $\mathbf {Z}$ : it transforms the matrices such that each dimension has unit variance and that the dimensions are uncorrelated. Intuitively, the vector spaces are easier to align along directions of high variance. S3) Dewhitening. A transformation inverse to S2: for improved performance it is important to restore the variance information after the projection, if whitening was applied in S2 BIBREF31 . S4) Symmetric re-weighting. This step attempts to further align the embeddings in the cross-lingual embedding space by measuring how well a dimension in the space correlates across languages for the current iteration dictionary $D^{(k)}$ . The best results are obtained when re-weighting is neutral to the projection direction, that is, when it is applied symmetrically in both languages. In the actual implementation S1 is applied only once, before self-learning. S2, S3 and S4 are applied in each self-learning iteration. Model Configurations. Note that C2 and C3 can be equally used on top of any (provided) seed lexicon (i.e., $D^{(1)}$ := $D_0$ ) to enable weakly supervised learning, as we propose here. In fact, the variations of the three key components, C1) seed lexicon, C2) self-learning, and C3) preprocessing and postprocessing, construct various model configurations which can be analyzed to probe the importance of each component in the CLWE induction process. A selection of representative configurations evaluated later in § "Results and Discussion" is summarized in Table 1 . ### Experimental Setup
Evaluation Task. Our task is bilingual lexicon induction (BLI). It has become the de facto standard evaluation for projection-based CLWEs BIBREF16 , BIBREF17 . In short, after a shared CLWE space has been induced, the task is to retrieve target language translations for a test set of source language words. Its lightweight nature allows us to conduct a comprehensive evaluation across a large number of language pairs. Since BLI is cast as a ranking task, following glavas2019howto we use mean average precision (MAP) as the main evaluation metric: in our BLI setup with only one correct translation for each “query” word, MAP is equal to mean reciprocal rank (MRR). (Selection of) Language Pairs. Our selection of test languages is guided by the following goals: a) following recent initiatives in other NLP research (e.g., for language modeling) BIBREF39 , BIBREF40 , we aim to ensure the coverage of different genealogical and typological language properties, and b) we aim to analyze a large set of language pairs and offer new evaluation data which extends and surpasses other work in the CLWE literature. These two properties will facilitate analyses between (dis)similar language pairs and offer a comprehensive set of evaluation setups that test the robustness and portability of fully unsupervised CLWEs. The final list of 15 diverse test languages is provided in Table 2 , and includes samples from different languages types and families. We run BLI evaluations for all language pairs in both directions, for a total of 15 $\times $ 14=210 BLI setups. Monolingual Embeddings. We use the 300-dim vectors of Grave:2018lrec for all 15 languages, pretrained on Common Crawl and Wikipedia with fastText BIBREF41 . We trim all vocabularies to the 200K most frequent words. Training and Test Dictionaries. They are derived from PanLex BIBREF43 , BIBREF44 , which was used in prior work on cross-lingual word embeddings BIBREF45 , BIBREF46 . PanLex currently spans around 1,300 language varieties with over 12M expressions: it offers some support and supervision also for low-resource language pairs BIBREF47 . For each source language ( $L_1$ ), we automatically translate their vocabulary words (if they are present in PanLex) to all 14 target ( $L_2$ ) languages. To ensure the reliability of the translation pairs, we retain only unigrams found in the vocabularies of the respective $L_2$ monolingual spaces which scored above a PanLex-predefined threshold. As in prior work BIBREF23 , BIBREF17 , we then reserve the 5K pairs created from the more frequent $L_1$ words for training, while the next 2K pairs are used for test. Smaller training dictionaries (1K and 500 pairs) are created by again selecting pairs comprising the most frequent $L_1$ words. Training Setup. In all experiments, we set the hyper-parameters to values that were tuned in prior research. When extracting the unsupervised seed lexicon, the 4K most frequent words of each language are used; self-learning operates on the 20K most frequent words of each language; with dropout the keep probability $p$ is 0.1; CSLS with $k=10$ nearest neighbors BIBREF24 . Again, Table 1 lists the main model configurations in our comparison. For the fully unsupervised model we always report the best performing configuration after probing different self-learning strategies (i.e., +sl, +sl+nod, and +sl+sym are tested). The results for unsupervised are always reported as averages over 5 restarts: this means that with unsupervised we count BLI setups as unsuccessful only if the results are close to zero in all 5/5 runs. orthg-super is the standard supervised model with orthogonal projections from prior work BIBREF21 , BIBREF17 . ### Results and Discussion
Main BLI results averaged over each source language ( $L_1$ ) are provided in Table 3 and Table 4 . We now summarize and discuss the main findings across several dimensions of comparison. Unsupervised vs. (Weakly) Supervised. First, when exactly the same components C2 and C3 are used, unsupervised is unable to outperform a (weakly) supervised full+sl+sym variant, and the gap in final performance is often substantial. In fact, full+sl+sym and full+sl+nod outperform the best unsupervised for all 210/210 BLI setups: we observe the same phenomenon with varying dictionary sizes, that is, it equally holds when we seed self-learning with 5K, 1K, and 500 translation pairs, see also Figure 2 . This also suggests that the main reason why unsupervised approaches were considered on-par with supervised approaches in prior work BIBREF23 , BIBREF24 is because they were not compared under fair circumstances: while unsupervised relied heavily on the components C2 and C3, these were omitted when running supervised baselines. Our unbiased comparison reveals that there is a huge gap even when supervised projection-based approaches consume only several hundred translation pairs to initiate self-learning. Are Unsupervised CLWEs Robust? The results also indicate that, contrary to the beliefs established by very recent work BIBREF24 , BIBREF30 , fully unsupervised approaches are still prone to getting stuck in local optima, and still suffer from robustness issues when dealing with distant language pairs: 87 out of 210 BLI setups ( $=41.4\%$ ) result in (near-)zero BLI performance, see also Table 4 . At the same time, weakly supervised methods with a seed lexicon of 1k or 500 pairs do not suffer from the robustness problem and always converge to a good solution, as also illustrated by the results reported in Table 5 . How Important are Preprocessing and Postprocessing? The comparisons between orthg-super (and orthg+sl+sym) on the one hand, and full-super (and full+sl+sym) on the other hand clearly indicate that the component C3 plays a substantial role in effective CLWE learning. full-super, which employs all steps S1-S4 (see § "Methodology" ), outperforms orthg-super in 208/210 setups with $|D_0|$ =5k and in 210/210 setups with $|D_0|$ =1k. Similarly, full+sl+sym is better than orthg+sl+sym in 210/210 setups (both for $|D_0|$ =1k,5k). The scores also indicate that dropout with self-learning is useful only when we work with noisy unsupervised seed lexicons: full+sl+nod and full+sl+sym without dropout consistently outperform full+sl across the board. How Important is (Robust) Self-Learning? We note that the best self-learning method is often useful even when $|D_0|=5k$ (i.e., full+sl+sym is better than full-super in 164/210 setups). However, the importance of robust self-learning gets more pronounced as we decrease the size of $D_0$ : full+sl+sym is better than full-super in 210/210 setups when $|D_0|=500$ or $|D_0|=1,000$ . The gap between the two models, as shown in Figure 2 , increases dramatically in favor of full+sl+sym as we decrease $|D_0|$ . Again, just comparing full-super and unsupervised in Figure 2 might give a false impression that fully unsupervised CLWE methods can match their supervised counterparts, but the comparison to full+sl+sym reveals the true extent of performance drop when we abandon even weak supervision. The scores also reveal that the choice of self-learning (C2) does matter: all best performing BLI runs with $|D_0|=1k$ are obtained by two configs with self-learning, and full+sl+sym is the best configuration for 177/210 setups (see Table 4 ). Language Pairs. As suggested before by sogaard2018on and further verified by glavas2019howto and doval2019onthe, the language pair at hand can have a huge impact on CLWE induction: the adversarial method of conneau2018word often gets stuck in poor local optima and yields degenerate solutions for distant language pairs such as English-Finnish. More recent CLWE methods BIBREF24 , BIBREF30 focus on mitigating this robustness issue. However, they still rely on one critical assumption which leads them to degraded performance for distant language pairs: they assume approximate isomorphism BIBREF49 , BIBREF48 between monolingual embedding spaces to learn the initial seed dictionary. In other words, they assume very similar geometric constellations between two monolingual spaces: due to the Zipfian phenomena in language BIBREF50 such near-isomorphism can be satisfied only for similar languages and for similar domains used for training monolingual vectors. This property is reflected in the results reported in Table 3 , the number of unsuccessful setups in Table 4 , as well as later in Figure 4 . For instance, the largest number of unsuccessful BLI setups with the unsupervised model is reported for Korean, Thai (a tonal language), and Basque (a language isolate): their morphological and genealogical properties are furthest away from other languages in our comparison. A substantial number of unsuccessful setups is also observed with other two language outliers from our set (see Table 2 again), Georgian and Indonesian, as well as with morphologically-rich languages such as Estonian or Turkish. One setting in which fully unsupervised methods did show impressive results in prior work are similar language pairs. However, even in these settings when the comparison to the weakly supervised full-super+sym is completely fair (i.e., the same components C2 and C3 are used for both), unsupervised still falls short of full-super+sym. These results for three source languages are summarized in Figure 3 . What is more, one could argue that we do not need unsupervised CLWEs for similar languages in the first place: we can harvest cheap supervision here, e.g., cognates. The main motivation behind unsupervised approaches is to support dissimilar and resource-poor language pairs for which supervision cannot be guaranteed. Domain Differences. Finally, we also verify that unsupervised CLWEs still cannot account for domain differences when training monolingual vectors. We rely on the probing test of sogaard2018on: 300-dim fastText vectors are trained on 1.1M sentences on three corpora: 1) EuroParl.v7 BIBREF51 (parliamentary proceedings); 2) Wikipedia BIBREF52 , and 3) EMEA BIBREF53 (medical), and BLI evaluation for three language pairs is conducted on standard MUSE BLI test sets BIBREF23 . The results, summarized in Figure 4 , reveal that unsupervised methods are able to yield a good solution only when there is no domain mismatch and for the pair with two most similar languages (English-Spanish), again questioning their robustness and portability to truly low-resource and more challenging setups. Weakly supervised methods ( $|D_0|=500$ or $D_0$ seeded with identical strings), in contrast, yield good solutions for all setups. ### Further Discussion and Conclusion
The superiority of weakly supervised methods (e.g., full+sl+sym) over unsupervised methods is especially pronounced for distant and typologically heterogeneous language pairs. However, our study also indicates that even carefully engineered projection-based methods with some seed supervision yield lower absolute performance for such pairs. While we have witnessed the proliferation of fully unsupervised CLWE models recently, some fundamental questions still remain. For instance, the underlying assumption of all projection-based methods (both supervised and unsupervised) is the topological similarity between monolingual spaces, which is why standard simple linear projections result in lower absolute BLI scores for distant pairs (see Table 4 and results in the supplemental material). Unsupervised approaches even exploit the assumption twice as their seed extraction is fully based on the topological similarity. Future work should move beyond the restrictive assumption by exploring new methods that can, e.g., 1) increase the isomorphism between monolingual spaces BIBREF54 by distinguishing between language-specific and language-pair-invariant subspaces; 2) learn effective non-linear or multiple local projections between monolingual spaces similar to the preliminary work of nakashole2018norma; 3) similar to vulic2016on and Lubin:2019naacl “denoisify” seed lexicons during the self-learning procedure. For instance, keeping only mutual/symmetric nearest neighbour as in full+sl+sym can be seen as a form of rudimentary denoisifying: it is indicative to see that the best overall performance in this work is reported with that model configuration. Further, the most important contributions of unsupervised CLWE models are, in fact, the improved and more robust self-learning procedures (component C2) and technical enhancements (component C3). In this work we have demonstrated that these components can be equally applied to weakly supervised approaches: starting from a set of only several hundred pairs, they can guarantee consistently improved performance across the board. As there is still no clear-cut use case scenario for unsupervised CLWEs, instead of “going fully unsupervised”, one pragmatic approach to widen the scope of CLWE learning and its application might invest more effort into extracting at least some seed supervision for a variety of language pairs BIBREF22 . This finding aligns well with the ongoing initiatives of the PanLex project BIBREF44 and the ASJP database BIBREF56 , which aim to collate at least some translation pairs in most of the world’s languages. Finally, this paper demonstrates that, in order to enable fair comparisons, future work on CLWEs should focus on evaluating the CLWE methods' constituent components (e.g, components C1-C3 from this work) instead of full-blown composite systems directly. One goal of the paper is to acknowledge that the work on fully unsupervised CLWE methods has indeed advanced state-of-the-art in cross-lingual word representation learning by offering new solutions also to weakly supervised CLWE methods. However, the robustness problems are still prominent with fully unsupervised CLWEs, and future work should invest more time and effort into developing more robust and more effective methods, e.g., by reaching beyond projection-based methods towards joint approaches BIBREF16 , BIBREF57 . ### Acknowledgments
This work is supported by the ERC Consolidator Grant LEXICAL: Lexical Acquisition Across Languages (no 648909). The work of Goran Glavaš is supported by the Baden-Württemberg Stiftung (AGREE grant of the Eliteprogramm). Roi Reichart is partially funded by ISF personal grants No. 1625/18. We thank the three anonymous reviewers for their encouraging comments and suggestions. Figure 1: General unsupervised CLWE approach. Table 1: Configurations obtained by varying components C1, C2, and C3 used in our empirical comparison in §4. Table 2: The list of 15 languages from our main BLI experiments along with their corresponding language family (IE = Indo-European), broad morphological type, and their ISO 639-1 code. Table 3: BLI scores (MRR) for all model configurations. The scores are averaged over all experimental setups where each of the 15 languages is used as L1: e.g., CA-* means that the translation direction is from Catalan (CA) as source (L1) to each of the remaining 14 languages listed in Table 2 as targets (L2), and we average over the corresponding 14 CA-* BLI setups. 5k and 1k denote the seed dictionary size for (weakly) supervised methods (D0). Unsuccessful setups refer to the number of BLI experimental setups with the fully UNSUPERVISED model that yield an MRR score ≤ 0.01. The Avg column refers to the averaged MRR scores of each model configuration over all 15×14=210 BLI setups. The highest scores for two different seed dictionary sizes in each column are in bold, the second best are underlined. See Table 1 for the brief description of all model configurations in the comparison. Full results for each particular language pair are available in the supplemental material. Table 4: Summary statistics computed over all 15×14=210 BLI setups. a) Unsuc. denotes the total number of unsuccessful setups, where a setup is considered unsuccessful if MRR ≤ 0.01 or MRR ≤ 0.05 (in the parentheses); b) Win refers to the total number of “winning” setups, that is, for all language pairs it counts how many times each particular model yields the best overall MRR score. We compute separate statistics for two settings (|D0| = 1k and |D0| = 5k). Figure 2: A comparison of average BLI scores with different seed dictionary sizes D0 between a fully unsupervised method (UNSUPER), a supervised method without self-learning (SUPER), and two best performing weakly supervised methods with self learning (+SL+NOD and +SL+SYM). While SUPER without selflearning displays a steep drop in performance with smaller seed dictionaries, there is only a slight decrease when self-learning is used: e.g., 500 translation pairs are still sufficient to initialize robust self-learning. Table 5: Results for a selection of BLI setups which were unsuccessful with the UNSUPERVIED CLWE method. Figure 3: A comparison of BLI scores on “easy” (i.e., similar) language pairs between the fully UNSUPERVISED model and a weakly supervised model (seed dictionary size |D0| = 200 or |D0| = 500) which relies on the self-learning procedure with the symmetry constraint (FULL+SL+SYM). Figure 4: BLI scores with the (most robust) fully UNSUPERVISED model for different language pairs when monolingual word embeddings are trained on dissimilar domains: parliamentary proceedings (EuroParl), Wikipedia (Wiki), and medical corpora (EMEA). Training and test data are the same as in (Søgaard et al., 2018). Table 6: All BLI scores (MRR) with Bulgarian (BG) as the source language. 5k and 1k denote the seed dictionary D0 size for (weakly) supervised methods. See Table 1 for a brief description of each model configuration. Table 7: All BLI scores (MRR) with Catalan (CA) as the source language. 5k and 1k denote the seed dictionary D0 size for (weakly) supervised methods. See Table 1 for a brief description of each model configuration. Table 8: All BLI scores (MRR) with Esperanto (EO) as the source language. 5k and 1k denote the seed dictionary D0 size for (weakly) supervised methods. See Table 1 for a brief description of each model configuration. Table 9: All BLI scores (MRR) with Estonian (ET) as the source language. 5k and 1k denote the seed dictionary D0 size for (weakly) supervised methods. See Table 1 for a brief description of each model configuration. Table 10: All BLI scores (MRR) with Basque (EU) as the source language. 5k and 1k denote the seed dictionary D0 size for (weakly) supervised methods. See Table 1 for a brief description of each model configuration. Table 11: All BLI scores (MRR) with Finnish (FI) as the source language. 5k and 1k denote the seed dictionary D0 size for (weakly) supervised methods. See Table 1 for a brief description of each model configuration. Table 12: All BLI scores (MRR) with Hebrew (HE) as the source language. 5k and 1k denote the seed dictionary D0 size for (weakly) supervised methods. See Table 1 for a brief description of each model configuration. Table 13: All BLI scores (MRR) with Hungarian (HU) as the source language. 5k and 1k denote the seed dictionary D0 size for (weakly) supervised methods. See Table 1 for a brief description of each model configuration. Table 14: All BLI scores (MRR) with Indonesian (ID) as the source language. 5k and 1k denote the seed dictionary D0 size for (weakly) supervised methods. See Table 1 for a brief description of each model configuration. Table 15: All BLI scores (MRR) with Georgian (KA) as the source language. 5k and 1k denote the seed dictionary D0 size for (weakly) supervised methods. See Table 1 for a brief description of each model configuration. Table 16: All BLI scores (MRR) with Korean (KO) as the source language. 5k and 1k denote the seed dictionary D0 size for (weakly) supervised methods. See Table 1 for a brief description of each model configuration. Table 17: All BLI scores (MRR) with Lithuanian (LT) as the source language. 5k and 1k denote the seed dictionary D0 size for (weakly) supervised methods. See Table 1 for a brief description of each model configuration. Table 18: All BLI scores (MRR) with Norwegian (NO) as the source language. 5k and 1k denote the seed dictionary D0 size for (weakly) supervised methods. See Table 1 for a brief description of each model configuration. Table 19: All BLI scores (MRR) with Thai (TH) as the source language. 5k and 1k denote the seed dictionary D0 size for (weakly) supervised methods. See Table 1 for a brief description of each model configuration. Table 20: All BLI scores (MRR) with Turkish (TR) as the source language. 5k and 1k denote the seed dictionary D0 size for (weakly) supervised methods. See Table 1 for a brief description of each model configuration. | the latest CLWE developments almost exclusively focus on fully unsupervised approaches BIBREF23 , BIBREF24 , BIBREF25 , BIBREF26 , BIBREF27 , BIBREF28 , BIBREF29 , BIBREF30 : they fully abandon any source of (even weak) supervision and extract the initial seed dictionary by exploiting topological similarities between pre-trained monolingual embedding spaces |
What is the significance of Sandra persuading her paper into letting her write human interest stories? How does this affect the text’s composition?
A. The human interest stories—i.e., Sandra’s interviews—provide the story’s central irony. The fact that humans cannot defeat the machine shows that the real interest is not human, but robotic.
B. The human interest stories provide a structure for the story to sit on. As she watches each player challenge the machine, it becomes more and more apparent that human personality cannot win.
C. -The human interest stories provide a structure for the story to sit on. As Doc introduces her to each chess player, their backstories help to unpack the significance of the chess tournament.
D. The human interest stories—i.e., Sandra’s interviews—provide a red herring for the story’s central goal, which is to hide the fact of Dr. Krakatower’s ability to beat the WBM machine.
| THE 64-SQUARE MADHOUSE by FRITZ LEIBER The machine was not perfect. It could be tricked. It could make mistakes. And—it could learn! [Transcriber's Note: This etext was produced from Worlds of If Science Fiction, May 1962. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed.] Silently, so as not to shock anyone with illusions about well dressed young women, Sandra Lea Grayling cursed the day she had persuaded the Chicago Space Mirror that there would be all sorts of human interest stories to be picked up at the first international grandmaster chess tournament in which an electronic computing machine was entered. Not that there weren't enough humans around, it was the interest that was in doubt. The large hall was crammed with energetic dark-suited men of whom a disproportionately large number were bald, wore glasses, were faintly untidy and indefinably shabby, had Slavic or Scandinavian features, and talked foreign languages. They yakked interminably. The only ones who didn't were scurrying individuals with the eager-zombie look of officials. Chess sets were everywhere—big ones on tables, still bigger diagram-type electric ones on walls, small peg-in sets dragged from side pockets and manipulated rapidly as part of the conversational ritual and still smaller folding sets in which the pieces were the tiny magnetized disks used for playing in free-fall. There were signs featuring largely mysterious combinations of letters: FIDE, WBM, USCF, USSF, USSR and UNESCO. Sandra felt fairly sure about the last three. The many clocks, bedside table size, would have struck a familiar note except that they had little red flags and wheels sprinkled over their faces and they were all in pairs, two clocks to a case. That Siamese-twin clocks should be essential to a chess tournament struck Sandra as a particularly maddening circumstance. Her last assignment had been to interview the pilot pair riding the first American manned circum-lunar satellite—and the five alternate pairs who hadn't made the flight. This tournament hall seemed to Sandra much further out of the world. Overheard scraps of conversation in reasonably intelligible English were not particularly helpful. Samples: "They say the Machine has been programmed to play nothing but pure Barcza System and Indian Defenses—and the Dragon Formation if anyone pushes the King Pawn." "Hah! In that case...." "The Russians have come with ten trunkfuls of prepared variations and they'll gang up on the Machine at adjournments. What can one New Jersey computer do against four Russian grandmasters?" "I heard the Russians have been programmed—with hypnotic cramming and somno-briefing. Votbinnik had a nervous breakdown." "Why, the Machine hasn't even a Haupturnier or an intercollegiate won. It'll over its head be playing." "Yes, but maybe like Capa at San Sebastian or Morphy or Willie Angler at New York. The Russians will look like potzers." "Have you studied the scores of the match between Moon Base and Circum-Terra?" "Not worth the trouble. The play was feeble. Barely Expert Rating." Sandra's chief difficulty was that she knew absolutely nothing about the game of chess—a point that she had slid over in conferring with the powers at the Space Mirror , but that now had begun to weigh on her. How wonderful it would be, she dreamed, to walk out this minute, find a quiet bar and get pie-eyed in an evil, ladylike way. "Perhaps mademoiselle would welcome a drink?" "You're durn tootin' she would!" Sandra replied in a rush, and then looked down apprehensively at the person who had read her thoughts. It was a small sprightly elderly man who looked like a somewhat thinned down Peter Lorre—there was that same impression of the happy Slavic elf. What was left of his white hair was cut very short, making a silvery nap. His pince-nez had quite thick lenses. But in sharp contrast to the somberly clad men around them, he was wearing a pearl-gray suit of almost exactly the same shade as Sandra's—a circumstance that created for her the illusion that they were fellow conspirators. "Hey, wait a minute," she protested just the same. He had already taken her arm and was piloting her toward the nearest flight of low wide stairs. "How did you know I wanted a drink?" "I could see that mademoiselle was having difficulty swallowing," he replied, keeping them moving. "Pardon me for feasting my eyes on your lovely throat." "I didn't suppose they'd serve drinks here." "But of course." They were already mounting the stairs. "What would chess be without coffee or schnapps?" "Okay, lead on," Sandra said. "You're the doctor." "Doctor?" He smiled widely. "You know, I like being called that." "Then the name is yours as long as you want it—Doc." Meanwhile the happy little man had edged them into the first of a small cluster of tables, where a dark-suited jabbering trio was just rising. He snapped his fingers and hissed through his teeth. A white-aproned waiter materialized. "For myself black coffee," he said. "For mademoiselle rhine wine and seltzer?" "That'd go fine." Sandra leaned back. "Confidentially, Doc, I was having trouble swallowing ... well, just about everything here." He nodded. "You are not the first to be shocked and horrified by chess," he assured her. "It is a curse of the intellect. It is a game for lunatics—or else it creates them. But what brings a sane and beautiful young lady to this 64-square madhouse?" Sandra briefly told him her story and her predicament. By the time they were served, Doc had absorbed the one and assessed the other. "You have one great advantage," he told her. "You know nothing whatsoever of chess—so you will be able to write about it understandably for your readers." He swallowed half his demitasse and smacked his lips. "As for the Machine—you do know, I suppose, that it is not a humanoid metal robot, walking about clanking and squeaking like a late medieval knight in armor?" "Yes, Doc, but...." Sandra found difficulty in phrasing the question. "Wait." He lifted a finger. "I think I know what you're going to ask. You want to know why, if the Machine works at all, it doesn't work perfectly, so that it always wins and there is no contest. Right?" Sandra grinned and nodded. Doc's ability to interpret her mind was as comforting as the bubbly, mildly astringent mixture she was sipping. He removed his pince-nez, massaged the bridge of his nose and replaced them. "If you had," he said, "a billion computers all as fast as the Machine, it would take them all the time there ever will be in the universe just to play through all the possible games of chess, not to mention the time needed to classify those games into branching families of wins for White, wins for Black and draws, and the additional time required to trace out chains of key-moves leading always to wins. So the Machine can't play chess like God. What the Machine can do is examine all the likely lines of play for about eight moves ahead—that is, four moves each for White and Black—and then decide which is the best move on the basis of capturing enemy pieces, working toward checkmate, establishing a powerful central position and so on." "That sounds like the way a man would play a game," Sandra observed. "Look ahead a little way and try to make a plan. You know, like getting out trumps in bridge or setting up a finesse." "Exactly!" Doc beamed at her approvingly. "The Machine is like a man. A rather peculiar and not exactly pleasant man. A man who always abides by sound principles, who is utterly incapable of flights of genius, but who never makes a mistake. You see, you are finding human interest already, even in the Machine." Sandra nodded. "Does a human chess player—a grandmaster, I mean—ever look eight moves ahead in a game?" "Most assuredly he does! In crucial situations, say where there's a chance of winning at once by trapping the enemy king, he examines many more moves ahead than that—thirty or forty even. The Machine is probably programmed to recognize such situations and do something of the same sort, though we can't be sure from the information World Business Machines has released. But in most chess positions the possibilities are so very nearly unlimited that even a grandmaster can only look a very few moves ahead and must rely on his judgment and experience and artistry. The equivalent of those in the Machine is the directions fed into it before it plays a game." "You mean the programming?" "Indeed yes! The programming is the crux of the problem of the chess-playing computer. The first practical model, reported by Bernstein and Roberts of IBM in 1958 and which looked four moves ahead, was programmed so that it had a greedy worried tendency to grab at enemy pieces and to retreat its own whenever they were attacked. It had a personality like that of a certain kind of chess-playing dub—a dull-brained woodpusher afraid to take the slightest risk of losing material—but a dub who could almost always beat an utter novice. The WBM machine here in the hall operates about a million times as fast. Don't ask me how, I'm no physicist, but it depends on the new transistors and something they call hypervelocity, which in turn depends on keeping parts of the Machine at a temperature near absolute zero. However, the result is that the Machine can see eight moves ahead and is capable of being programmed much more craftily." "A million times as fast as the first machine, you say, Doc? And yet it only sees twice as many moves ahead?" Sandra objected. "There is a geometrical progression involved there," he told her with a smile. "Believe me, eight moves ahead is a lot of moves when you remember that the Machine is errorlessly examining every one of thousands of variations. Flesh-and-blood chess masters have lost games by blunders they could have avoided by looking only one or two moves ahead. The Machine will make no such oversights. Once again, you see, you have the human factor, in this case working for the Machine." "Savilly, I have been looking allplace for you!" A stocky, bull-faced man with a great bristling shock of black, gray-flecked hair had halted abruptly by their table. He bent over Doc and began to whisper explosively in a guttural foreign tongue. Sandra's gaze traveled beyond the balustrade. Now that she could look down at it, the central hall seemed less confusedly crowded. In the middle, toward the far end, were five small tables spaced rather widely apart and with a chessboard and men and one of the Siamese clocks set out on each. To either side of the hall were tiers of temporary seats, about half of them occupied. There were at least as many more people still wandering about. On the far wall was a big electric scoreboard and also, above the corresponding tables, five large dully glassy chessboards, the White squares in light gray, the Black squares in dark. One of the five wall chessboards was considerably larger than the other four—the one above the Machine. Sandra looked with quickening interest at the console of the Machine—a bank of keys and some half-dozen panels of rows and rows of tiny telltale lights, all dark at the moment. A thick red velvet cord on little brass standards ran around the Machine at a distance of about ten feet. Inside the cord were only a few gray-smocked men. Two of them had just laid a black cable to the nearest chess table and were attaching it to the Siamese clock. Sandra tried to think of a being who always checked everything, but only within limits beyond which his thoughts never ventured, and who never made a mistake.... "Miss Grayling! May I present to you Igor Jandorf." She turned back quickly with a smile and a nod. "I should tell you, Igor," Doc continued, "that Miss Grayling represents a large and influential Midwestern newspaper. Perhaps you have a message for her readers." The shock-headed man's eyes flashed. "I most certainly do!" At that moment the waiter arrived with a second coffee and wine-and-seltzer. Jandorf seized Doc's new demitasse, drained it, set it back on the tray with a flourish and drew himself up. "Tell your readers, Miss Grayling," he proclaimed, fiercely arching his eyebrows at her and actually slapping his chest, "that I, Igor Jandorf, will defeat the Machine by the living force of my human personality! Already I have offered to play it an informal game blindfold—I, who have played 50 blindfold games simultaneously! Its owners refuse me. I have challenged it also to a few games of rapid-transit—an offer no true grandmaster would dare ignore. Again they refuse me. I predict that the Machine will play like a great oaf—at least against me . Repeat: I, Igor Jandorf, by the living force of my human personality, will defeat the Machine. Do you have that? You can remember it?" "Oh yes," Sandra assured him, "but there are some other questions I very much want to ask you, Mr. Jandorf." "I am sorry, Miss Grayling, but I must clear my mind now. In ten minutes they start the clocks." While Sandra arranged for an interview with Jandorf after the day's playing session, Doc reordered his coffee. "One expects it of Jandorf," he explained to Sandra with a philosophic shrug when the shock-headed man was gone. "At least he didn't take your wine-and-seltzer. Or did he? One tip I have for you: don't call a chess master Mister, call him Master. They all eat it up." "Gee, Doc, I don't know how to thank you for everything. I hope I haven't offended Mis—Master Jandorf so that he doesn't—" "Don't worry about that. Wild horses couldn't keep Jandorf away from a press interview. You know, his rapid-transit challenge was cunning. That's a minor variety of chess where each player gets only ten seconds to make a move. Which I don't suppose would give the Machine time to look three moves ahead. Chess players would say that the Machine has a very slow sight of the board. This tournament is being played at the usual international rate of 15 moves an hour, and—" "Is that why they've got all those crazy clocks?" Sandra interrupted. "Oh, yes. Chess clocks measure the time each player takes in making his moves. When a player makes a move he presses a button that shuts his clock off and turns his opponent's on. If a player uses too much time, he loses as surely as if he were checkmated. Now since the Machine will almost certainly be programmed to take an equal amount of time on successive moves, a rate of 15 moves an hour means it will have 4 minutes a move—and it will need every second of them! Incidentally it was typical Jandorf bravado to make a point of a blindfold challenge—just as if the Machine weren't playing blindfold itself. Or is the Machine blindfold? How do you think of it?" "Gosh, I don't know. Say, Doc, is it really true that Master Jandorf has played 50 games at once blindfolded? I can't believe that." "Of course not!" Doc assured her. "It was only 49 and he lost two of those and drew five. Jandorf always exaggerates. It's in his blood." "He's one of the Russians, isn't he?" Sandra asked. "Igor?" Doc chuckled. "Not exactly," he said gently. "He is originally a Pole and now he has Argentinian citizenship. You have a program, don't you?" Sandra started to hunt through her pocketbook, but just then two lists of names lit up on the big electric scoreboard. THE PLAYERS William Angler, USA Bela Grabo, Hungary Ivan Jal, USSR Igor Jandorf, Argentina Dr. S. Krakatower, France Vassily Lysmov, USSR The Machine, USA (programmed by Simon Great) Maxim Serek, USSR Moses Sherevsky, USA Mikhail Votbinnik, USSR Tournament Director : Dr. Jan Vanderhoef FIRST ROUND PAIRINGS Sherevsky vs. Serek Jal vs. Angler Jandorf vs. Votbinnik Lysmov vs. Krakatower Grabo vs. Machine "Cripes, Doc, they all sound like they were Russians," Sandra said after a bit. "Except this Willie Angler. Oh, he's the boy wonder, isn't he?" Doc nodded. "Not such a boy any longer, though. He's.... Well, speak of the Devil's children.... Miss Grayling, I have the honor of presenting to you the only grandmaster ever to have been ex-chess-champion of the United States while still technically a minor—Master William Augustus Angler." A tall, sharply-dressed young man with a hatchet face pressed the old man back into his chair. "How are you, Savvy, old boy old boy?" he demanded. "Still chasing the girls, I see." "Please, Willie, get off me." "Can't take it, huh?" Angler straightened up somewhat. "Hey waiter! Where's that chocolate malt? I don't want it next year. About that ex- , though. I was swindled, Savvy. I was robbed." "Willie!" Doc said with some asperity. "Miss Grayling is a journalist. She would like to have a statement from you as to how you will play against the Machine." Angler grinned and shook his head sadly. "Poor old Machine," he said. "I don't know why they take so much trouble polishing up that pile of tin just so that I can give it a hit in the head. I got a hatful of moves it'll burn out all its tubes trying to answer. And if it gets too fresh, how about you and me giving its low-temperature section the hotfoot, Savvy? The money WBM's putting up is okay, though. That first prize will just fit the big hole in my bank account." "I know you haven't the time now, Master Angler," Sandra said rapidly, "but if after the playing session you could grant me—" "Sorry, babe," Angler broke in with a wave of dismissal. "I'm dated up for two months in advance. Waiter! I'm here, not there!" And he went charging off. Doc and Sandra looked at each other and smiled. "Chess masters aren't exactly humble people, are they?" she said. Doc's smile became tinged with sad understanding. "You must excuse them, though," he said. "They really get so little recognition or recompense. This tournament is an exception. And it takes a great deal of ego to play greatly." "I suppose so. So World Business Machines is responsible for this tournament?" "Correct. Their advertising department is interested in the prestige. They want to score a point over their great rival." "But if the Machine plays badly it will be a black eye for them," Sandra pointed out. "True," Doc agreed thoughtfully. "WBM must feel very sure.... It's the prize money they've put up, of course, that's brought the world's greatest players here. Otherwise half of them would be holding off in the best temperamental-artist style. For chess players the prize money is fabulous—$35,000, with $15,000 for first place, and all expenses paid for all players. There's never been anything like it. Soviet Russia is the only country that has ever supported and rewarded her best chess players at all adequately. I think the Russian players are here because UNESCO and FIDE (that's Federation Internationale des Echecs —the international chess organization) are also backing the tournament. And perhaps because the Kremlin is hungry for a little prestige now that its space program is sagging." "But if a Russian doesn't take first place it will be a black eye for them." Doc frowned. "True, in a sense. They must feel very sure.... Here they are now." Four men were crossing the center of the hall, which was clearing, toward the tables at the other end. Doubtless they just happened to be going two by two in close formation, but it gave Sandra the feeling of a phalanx. "The first two are Lysmov and Votbinnik," Doc told her. "It isn't often that you see the current champion of the world—Votbinnik—and an ex-champion arm in arm. There are two other persons in the tournament who have held that honor—Jal and Vanderhoef the director, way back." "Will whoever wins this tournament become champion?" "Oh no. That's decided by two-player matches—a very long business—after elimination tournaments between leading contenders. This tournament is a round robin: each player plays one game with every other player. That means nine rounds." "Anyway there are an awful lot of Russians in the tournament," Sandra said, consulting her program. "Four out of ten have USSR after them. And Bela Grabo, Hungary—that's a satellite. And Sherevsky and Krakatower are Russian-sounding names." "The proportion of Soviet to American entries in the tournament represents pretty fairly the general difference in playing strength between the two countries," Doc said judiciously. "Chess mastery moves from land to land with the years. Way back it was the Moslems and the Hindus and Persians. Then Italy and Spain. A little over a hundred years ago it was France and England. Then Germany, Austria and the New World. Now it's Russia—including of course the Russians who have run away from Russia. But don't think there aren't a lot of good Anglo-Saxon types who are masters of the first water. In fact, there are a lot of them here around us, though perhaps you don't think so. It's just that if you play a lot of chess you get to looking Russian. Once it probably made you look Italian. Do you see that short bald-headed man?" "You mean the one facing the Machine and talking to Jandorf?" "Yes. Now that's one with a lot of human interest. Moses Sherevsky. Been champion of the United States many times. A very strict Orthodox Jew. Can't play chess on Fridays or on Saturdays before sundown." He chuckled. "Why, there's even a story going around that one rabbi told Sherevsky it would be unlawful for him to play against the Machine because it is technically a golem —the clay Frankenstein's monster of Hebrew legend." Sandra asked, "What about Grabo and Krakatower?" Doc gave a short scornful laugh. "Krakatower! Don't pay any attention to him . A senile has-been, it's a scandal he's been allowed to play in this tournament! He must have pulled all sorts of strings. Told them that his lifelong services to chess had won him the honor and that they had to have a member of the so-called Old Guard. Maybe he even got down on his knees and cried—and all the time his eyes on that expense money and the last-place consolation prize! Yet dreaming schizophrenically of beating them all! Please, don't get me started on Dirty Old Krakatower." "Take it easy, Doc. He sounds like he would make an interesting article? Can you point him out to me?" "You can tell him by his long white beard with coffee stains. I don't see it anywhere, though. Perhaps he's shaved it off for the occasion. It would be like that antique womanizer to develop senile delusions of youthfulness." "And Grabo?" Sandra pressed, suppressing a smile at the intensity of Doc's animosity. Doc's eyes grew thoughtful. "About Bela Grabo (why are three out of four Hungarians named Bela?) I will tell you only this: That he is a very brilliant player and that the Machine is very lucky to have drawn him as its first opponent." He would not amplify his statement. Sandra studied the Scoreboard again. "This Simon Great who's down as programming the Machine. He's a famous physicist, I suppose?" "By no means. That was the trouble with some of the early chess-playing machines—they were programmed by scientists. No, Simon Great is a psychologist who at one time was a leading contender for the world's chess championship. I think WBM was surprisingly shrewd to pick him for the programming job. Let me tell you—No, better yet—" Doc shot to his feet, stretched an arm on high and called out sharply, "Simon!" A man some four tables away waved back and a moment later came over. "What is it, Savilly?" he asked. "There's hardly any time, you know." The newcomer was of middle height, compact of figure and feature, with graying hair cut short and combed sharply back. Doc spoke his piece for Sandra. Simon Great smiled thinly. "Sorry," he said, "But I am making no predictions and we are giving out no advance information on the programming of the Machine. As you know, I have had to fight the Players' Committee tooth and nail on all sorts of points about that and they have won most of them. I am not permitted to re-program the Machine at adjournments—only between games (I did insist on that and get it!) And if the Machine breaks down during a game, its clock keeps running on it. My men are permitted to make repairs—if they can work fast enough." "That makes it very tough on you," Sandra put in. "The Machine isn't allowed any weaknesses." Great nodded soberly. "And now I must go. They've almost finished the count-down, as one of my technicians keeps on calling it. Very pleased to have met you, Miss Grayling—I'll check with our PR man on that interview. Be seeing you, Savvy." The tiers of seats were filled now and the central space almost clear. Officials were shooing off a few knots of lingerers. Several of the grandmasters, including all four Russians, were seated at their tables. Press and company cameras were flashing. The four smaller wallboards lit up with the pieces in the opening position—white for White and red for Black. Simon Great stepped over the red velvet cord and more flash bulbs went off. "You know, Doc," Sandra said, "I'm a dog to suggest this, but what if this whole thing were a big fake? What if Simon Great were really playing the Machine's moves? There would surely be some way for his electricians to rig—" Doc laughed happily—and so loudly that some people at the adjoining tables frowned. "Miss Grayling, that is a wonderful idea! I will probably steal it for a short story. I still manage to write and place a few in England. No, I do not think that is at all likely. WBM would never risk such a fraud. Great is completely out of practice for actual tournament play, though not for chess-thinking. The difference in style between a computer and a man would be evident to any expert. Great's own style is remembered and would be recognized—though, come to think of it, his style was often described as being machinelike...." For a moment Doc's eyes became thoughtful. Then he smiled again. "But no, the idea is impossible. Vanderhoef as Tournament Director has played two or three games with the Machine to assure himself that it operates legitimately and has grandmaster skill." "Did the Machine beat him?" Sandra asked. Doc shrugged. "The scores weren't released. It was very hush-hush. But about your idea, Miss Grayling—did you ever read about Maelzel's famous chess-playing automaton of the 19th Century? That one too was supposed to work by machinery (cogs and gears, not electricity) but actually it had a man hidden inside it—your Edgar Poe exposed the fraud in a famous article. In my story I think the chess robot will break down while it is being demonstrated to a millionaire purchaser and the young inventor will have to win its game for it to cover up and swing the deal. Only the millionaire's daughter, who is really a better player than either of them ... yes, yes! Your Ambrose Bierce too wrote a story about a chess-playing robot of the clickety-clank-grr kind who murdered his creator, crushing him like an iron grizzly bear when the man won a game from him. Tell me, Miss Grayling, do you find yourself imagining this Machine putting out angry tendrils to strangle its opponents, or beaming rays of death and hypnotism at them? I can imagine...." While Doc chattered happily on about chess-playing robots and chess stories, Sandra found herself thinking about him. A writer of some sort evidently and a terrific chess buff. Perhaps he was an actual medical doctor. She'd read something about two or three coming over with the Russian squad. But Doc certainly didn't sound like a Soviet citizen. He was older than she'd first assumed. She could see that now that she was listening to him less and looking at him more. Tired, too. Only his dark-circled eyes shone with unquenchable youth. A useful old guy, whoever he was. An hour ago she'd been sure she was going to muff this assignment completely and now she had it laid out cold. For the umpteenth time in her career Sandra shied away from the guilty thought that she wasn't a writer at all or even a reporter, she just used dime-a-dozen female attractiveness to rope a susceptible man (young, old, American, Russian) and pick his brain.... She realized suddenly that the whole hall had become very quiet. Doc was the only person still talking and people were again looking at them disapprovingly. All five wallboards were lit up and the changed position of a few pieces showed that opening moves had been made on four of them, including the Machine's. The central space between the tiers of seats was completely clear now, except for one man hurrying across it in their direction with the rapid yet quiet, almost tip-toe walk that seemed to mark all the officials. Like morticians' assistants , she thought. He rapidly mounted the stairs and halted at the top to look around searchingly. His gaze lighted on their table, his eyebrows went up, and he made a beeline for Doc. Sandra wondered if she should warn him that he was about to be shushed. The official laid a hand on Doc's shoulder. "Sir!" he said agitatedly. "Do you realize that they've started your clock, Dr. Krakatower?" | C. -The human interest stories provide a structure for the story to sit on. As Doc introduces her to each chess player, their backstories help to unpack the significance of the chess tournament. |
How much does the standard metrics for style accuracy vary on different re-runs? | ### Introduction
Deep generative models attract a lot of attention in recent years BIBREF0. Such methods as variational autoencoders BIBREF1 or generative adversarial networks BIBREF2 are successfully applied to a variety of machine vision problems including image generation BIBREF3, learning interpretable image representations BIBREF4 and style transfer for images BIBREF5. However, natural language generation is more challenging due to many reasons, such as the discrete nature of textual information BIBREF6, the absence of local information continuity and non-smooth disentangled representations BIBREF7. Due to these difficulties, text generation is mostly limited to specific narrow applications and is usually working in supervised settings. Content and style are deeply fused in natural language, but style transfer for texts is often addressed in the context of disentangled latent representations BIBREF6, BIBREF8, BIBREF9, BIBREF10, BIBREF11, BIBREF12. Intuitive understanding of this problem is apparent: if an input text has some attribute $A$, a system generates new text similar to the input on a given set of attributes with only one attribute $A$ changed to the target attribute $\tilde{A}$. In the majority of previous works, style transfer is obtained through an encoder-decoder architecture with one or multiple style discriminators to learn disentangled representations. The encoder takes a sentence as an input and generates a style-independent content representation. The decoder then takes the content representation and the target style representation to generate the transformed sentence. In BIBREF13 authors question the quality and usability of the disentangled representations for texts and suggest an end-to-end approach to style transfer similar to an end-to-end machine translation. Contribution of this paper is three-fold: 1) we show that different style transfer architectures have varying results on test and that reporting error margins for various training re-runs of the same model is especially important for adequate assessment of the models accuracy, see Figure FIGREF1; 2) we show that BLEU BIBREF14 between input and output and accuracy of style transfer measured in terms of the accuracy of a pre-trained external style classifier can be manipulated and naturally diverge from the intuitive goal of the style transfer task starting from a certain threshold; 3) new architectures that perform style transfer using improved latent representations are shown to outperform state of the art in terms of BLEU between output and human-written reformulations. ### Related Work
Style of a text is a very general notion that is hard to define in rigorous terms BIBREF15. However, the style of a text can be characterized quantitatively BIBREF16; stylized texts could be generated if a system is trained on a dataset of stylistically similar texts BIBREF17; and author-style could be learned end-to-end BIBREF18, BIBREF19, BIBREF20. A majority of recent works on style transfer focus on the sentiment of text and use it as a target attribute. For example, in BIBREF21, BIBREF22, BIBREF23 estimate the quality of the style transfer with binary sentiment classifier trained on the corpora further used for the training of the style-transfer system. BIBREF24 and especially BIBREF9 generalize this ad-hoc approach defining a style as a set of arbitrary quantitively measurable categorial or continuous parameters. Such parameters could include the 'style of the time' BIBREF16, author-specific attributes (see BIBREF25 or BIBREF26 on 'shakespearization'), politeness BIBREF27, formality of speech BIBREF28, and gender or even political slant BIBREF29. A significant challenge associated with narrowly defined style-transfer problems is that finding a good solution for one aspect of a style does not guarantee that you can use the same solution for a different aspect of it. For example, BIBREF30 build a generative model for sentiment transfer with a retrieve-edit approach. In BIBREF21 a delete-retrieve model shows good results for sentiment transfer. However, it is hard to imagine that these retrieval approaches could be used, say, for the style of the time or formality, since in these cases the system is often expected to paraphrase a given sentence to achieve the target style. In BIBREF6 the authors propose a more general approach to the controlled text generation combining variational autoencoder (VAE) with an extended wake-sleep mechanism in which the sleep procedure updates both the generator and external classifier that assesses generated samples and feedbacks learning signals to the generator. Authors had concatenated labels for style with the text representation of the encoder and used this vector with "hard-coded" information about the sentiment of the output as the input of the decoder. This approach seems promising, and some other papers either extend it or use similar ideas. BIBREF8 applied a GAN to align the hidden representations of sentences from two corpora using an adversarial loss to decompose information about the form. In BIBREF31 model learns a smooth code space and can be used as a discrete GAN with the ability to generate coherent discrete outputs from continuous samples. Authors use two different generators for two different styles. In BIBREF9 an adversarial network is used to make sure that the output of the encoder does not have style representation. BIBREF6 also uses an adversarial component that ensures there is no stylistic information within the representation. BIBREF9 do not use a dedicated component that controls the semantic component of the latent representation. Such a component is proposed by BIBREF10 who demonstrate that decomposition of style and content could be improved with an auxiliary multi-task for label prediction and adversarial objective for bag-of-words prediction. BIBREF11 also introduces a dedicated component to control semantic aspects of latent representations and an adversarial-motivational training that includes a special motivational loss to encourage a better decomposition. Speaking about preservation of semantics one also has to mention works on paraphrase systems, see, for example BIBREF32, BIBREF33, BIBREF34. The methodology described in this paper could be extended to paraphrasing systems in terms of semantic preservation measurement, however, this is the matter of future work. BIBREF13 state that learning a latent representation, which is independent of the attributes specifying its style, is rarely attainable. There are other works on style transfer that are based on the ideas of neural machine translation with BIBREF35 and without parallel corpora BIBREF36 in line with BIBREF37 and BIBREF38. It is important to underline here that majority of the papers dedicated to style transfer for texts treat sentiment of a sentence as a stylistic rather than semantic attribute despite particular concerns BIBREF39. It is also crucial to mention that in line with BIBREF9 majority of the state of the art methods for style transfer use an external pre-trained classifier to measure the accuracy of the style transfer. BLEU computes the harmonic mean of precision of exact matching n-grams between a reference and a target sentence across the corpus. It is not sensitive to minute changes, but BLEU between input and output is often used as the coarse measure of the semantics preservation. For the corpora that have human written reformulations, BLEU between the output of the model and human text is used. These metrics are used alongside with a handful of others such as PINC (Paraphrase In N-gram Changes) score BIBREF35, POS distance BIBREF12, language fluency BIBREF10, etc. Figure FIGREF2 shows self-reported results of different models in terms of two most frequently measured performance metrics, namely, BLEU and Accuracy of the style transfer. This paper focuses on Yelp! reviews dataset that was lately enhanced with human written reformulations by BIBREF21. These are Yelp! reviews, where each short English review of a place is labeled as a negative or as a positive once. This paper studies three metrics that are most common in the field at the moment and questions to which extent can they be used for the performance assessment. These metrics are the accuracy of an external style classifier that is trained to measure the accuracy of the style transfer, BLEU between input and output of a system, and BLEU between output and human-written texts. ### Style transfer
In this work we experiment with extensions of a model, described in BIBREF6, using Texar BIBREF40 framework. To generate plausible sentences with specific semantic and stylistic features every sentence is conditioned on a representation vector $z$ which is concatenated with a particular code $c$ that specifies desired attribute, see Figure FIGREF8. Under notation introduced in BIBREF6 the base autoencoder (AE) includes a conditional probabilistic encoder $E$ defined with parameters $\theta _E$ to infer the latent representation $z$ given input $x$ Generator $G$ defined with parameters $\theta _G$ is a GRU-RNN for generating and output $\hat{x}$ defined as a sequence of tokens $\hat{x} = {\hat{x}_1, ..., \hat{x}_T}$ conditioned on the latent representation $z$ and a stylistic component $c$ that are concatenated and give rise to a generative distribution These encoder and generator form an AE with the following loss This standard reconstruction loss that drives the generator to produce realistic sentences is combined with two additional losses. The first discriminator provides extra learning signals which enforce the generator to produce coherent attributes that match the structured code in $c$. Since it is impossible to propagate gradients from the discriminator through the discrete sample $\hat{x}$, we use a deterministic continuous approximation a "soft" generated sentence, denoted as $\tilde{G} = \tilde{G}_\tau (z, c)$ with "temperature" $\tau $ set to $\tau \rightarrow 0$ as training proceeds. The resulting “soft” generated sentence is fed into the discriminator to measure the fitness to the target attribute, leading to the following loss Finally, under the assumption that each structured attribute of generated sentences is controlled through the corresponding code in $c$ and is independent from $z$ one would like to control that other not explicitly modelled attributes do not entangle with $c$. This is addressed by the dedicated loss The training objective for the baseline, shown in Figure FIGREF8, is therefore a sum of the losses from Equations (DISPLAY_FORM4) – (DISPLAY_FORM6) defined as where $\lambda _c$ and $\lambda _z$ are balancing parameters. Let us propose two further extensions of this baseline architecture. To improve reproducibility of the research the code of the studied models is open. Both extensions aim to improve the quality of information decomposition within the latent representation. In the first one, shown in Figure FIGREF12, a special dedicated discriminator is added to the model to control that the latent representation does not contain stylistic information. The loss of this discriminator is defined as Here a discriminator denoted as $D_z$ is trying to predict code $c$ using representation $z$. Combining the loss defined by Equation (DISPLAY_FORM7) with the adversarial component defined in Equation (DISPLAY_FORM10) the following learning objective is formed where $\mathcal {L}_{baseline}$ is a sum defined in Equation (DISPLAY_FORM7), $\lambda _{D_z}$ is a balancing parameter. The second extension of the baseline architecture does not use an adversarial component $D_z$ that is trying to eradicate information on $c$ from component $z$. Instead, the system, shown in Figure FIGREF16 feeds the "soft" generated sentence $\tilde{G}$ into encoder $E$ and checks how close is the representation $E(\tilde{G} )$ to the original representation $z = E(x)$ in terms of the cosine distance. We further refer to it as shifted autoencoder or SAE. Ideally, both $E(\tilde{G} (E(x), c))$ and $E(\tilde{G} (E(x), \bar{c}))$, where $\bar{c}$ denotes an inverse style code, should be both equal to $E(x)$. The loss of the shifted autoencoder is where $\lambda _{cos}$ and $\lambda _{cos^{-}}$ are two balancing parameters, with two additional terms in the loss, namely, cosine distances between the softened output processed by the encoder and the encoded original input, defined as We also study a combination of both approaches described above, shown on Figure FIGREF17. In Section SECREF4 we describe a series of experiments that we have carried out for these architectures using Yelp! reviews dataset. ### Experiments
We have found that the baseline, as well as the proposed extensions, have noisy outcomes, when retrained from scratch, see Figure FIGREF1. Most of the papers mentioned in Section SECREF2 measure the performance of the methods proposed for the sentiment transfer with two metrics: accuracy of the external sentiment classifier measured on test data, and BLEU between the input and output that is regarded as a coarse metric for semantic similarity. In the first part of this section, we demonstrate that reporting error margins is essential for the performance assessment in terms that are prevalent in the field at the moment, i.e., BLEU between input and output and accuracy of the external sentiment classifier. In the second part, we also show that both of these two metrics after a certain threshold start to diverge from an intuitive goal of the style transfer and could be manipulated. ### Experiments ::: Error margins matter
On Figure FIGREF1 one can see that the outcomes for every single rerun differ significantly. Namely, accuracy can change up to 5 percentage points, whereas BLEU can vary up to 8 points. This variance can be partially explained with the stochasticity incurred due to sampling from the latent variables. However, we show that results for state of the art models sometimes end up within error margins from one another, so one has to report the margins to compare the results rigorously. More importantly, one can see that there is an inherent trade-off between these two performance metrics. This trade-off is not only visible across models but is also present for the same retrained architecture. Therefore, improving one of the two metrics is not enough to confidently state that one system solves the style-transfer problem better than the other. One has to report error margins after several consecutive retrains and instead of comparing one of the two metrics has to talk about Pareto-like optimization that would show confident improvement of both. To put obtained results into perspective, we have retrained every model from scratch five times in a row. We have also retrained the models of BIBREF12 five times since their code is published online. Figure FIGREF19 shows the results of all models with error margins. It is also enhanced with other self-reported results on the same Yelp! review dataset for which no code was published. One can see that error margins of the models, for which several reruns could be performed, overlap significantly. In the next subsection, we carefully study BLEU and accuracy of the external classifier and discuss their aptness to measure style transfer performance. ### Experiments ::: Delete, duplicate and conquer
One can argue that as there is an inevitable entanglement between semantics and stylistics in natural language, there is also an apparent entanglement between BLEU of input and output and accuracy estimation of the style. Indeed, the output that copies input gives maximal BLEU yet clearly fails in terms of the style transfer. On the other hand, a wholly rephrased sentence could provide a low BLEU between input and output but high accuracy. These two issues are not problematic when both BLEU between input and output and accuracy of the transfer are relatively low. However, since style transfer methods have significantly evolved in recent years, some state of the art methods are now sensitive to these issues. The trade-off between these two metrics can be seen in Figure FIGREF1 as well as in Figure FIGREF19. As we have mentioned above, the accuracy of an external classifier and BLEU between output and input are the most widely used methods to assess the performance of style transfer at this moment. However, both of these metrics can be manipulated in a relatively simple manner. One can extend the generative architecture with internal pre-trained classifier of style and then perform the following heuristic procedure: measure the style accuracy on the output for a given batch; choose the sentences that style classifier labels as incorrect; replace them with duplicates of sentences from the given batch that have correct style according to the internal classifier and show the highest BLEU with given inputs. This way One can replace all sentences that push measured accuracy down and boost reported accuracy to 100%. To see the effect that this manipulation has on the key performance metric we split all sentences with wrong style in 10 groups of equal size and replaces them with the best possible duplicates of the stylistically correct sentences group after group. The results of this process are shown in Figure FIGREF24. This result is disconcerting. Simply replacing part of the output with duplicates of the sentences that happen to have relatively high BLEU with given inputs allows to "boost" accuracy to 100% and "improve" BLEU. The change of BLEU during such manipulation stays within error margins of the architecture, but accuracy is significantly manipulated. What is even more disturbing is that BLEU between such manipulated output of the batch and human-written reformulations provided in BIBREF12 also grows. Figure FIGREF24 shows that for SAE but all four architectures described in Section SECREF3 demonstrate similar behavior. Our experiments show that though we can manipulate BLEU between output and human-written text, it tends to change monotonically. That might be because of the fact that this metric incorporates information on stylistics and semantics of the text at the same time, preserving inevitable entanglement that we have mentioned earlier. Despite being costly, human-written reformulations are needed for future experiments with style transfer. It seems that modern architectures have reached a certain level of complexity for which naive proxy metrics such as accuracy of an external classifier or BLEU between output and input are already not enough for performance estimation and should be combined with BLEU between output and human-written texts. As the quality of style transfer grows further one has to improve the human-written data sets: for example, one would like to have data sets similar to the ones used for machine translation with several reformulations of the same sentence. On Figure FIGREF25 one can see how new proposed architectures compare with another state of the art approaches in terms of BLEU between output and human-written reformulations. ### Conclusion
Style transfer is not a rigorously defined NLP problem. Starting from definitions of style and semantics and finishing with metrics that could be used to evaluate the performance of a proposed system. There is a surge of recent contributions that work on this problem. This paper highlights several issues connected with this lack of rigor. First, it shows that the state of the art algorithms are inherently noisy on the two most widely accepted metrics, namely, BLEU between input and output and accuracy of the external style classifier. This noise can be partially attributed to the adversarial components that are often used in the state of the art architectures and partly due to certain methodological inconsistencies in the assessment of the performance. Second, it shows that reporting error margins of several consecutive retrains for the same model is crucial for the comparison of different architectures, since error margins for some of the models overlap significantly. Finally, it demonstrates that even BLEU on human-written reformulations can be manipulated in a relatively simple way. ### Supplemental Material
Here are some examples characteristic for different systems. An output of a system follows the input. Here are some successful examples produced by the system with additional discriminator: it's not much like an actual irish pub, which is depressing. $\rightarrow $ it's definitely much like an actual irish pub, which is grateful. i got a bagel breakfast sandwich and it was delicious! $\rightarrow $ i got a bagel breakfast sandwich and it was disgusting! i love their flavored coffee. $\rightarrow $ i dumb their flavored coffee. i got a bagel breakfast sandwich and it was delicious! $\rightarrow $ i got a bagel breakfast sandwich and it was disgusting! i love their flavored coffee. $\rightarrow $ i dumb their flavored coffee. nice selection of games to play. $\rightarrow $ typical selection of games to play. i'm not a fan of huge chain restaurants. $\rightarrow $ i'm definitely a fan of huge chain restaurants. Here are some examples of typical faulty reformulations: only now i'm really hungry, and really pissed off. $\rightarrow $ kids now i'm really hungry, and really extraordinary off. what a waste of my time and theirs. $\rightarrow $ what a wow. of my time and theirs. cooked to perfection and very flavorful. $\rightarrow $ cooked to pain and very outdated. the beer was nice and cold! $\rightarrow $ the beer was nice and consistant! corn bread was also good! $\rightarrow $ corn bread was also unethical bagged Here are some successful examples produced by the SAE: our waitress was the best, very accommodating. $\rightarrow $ our waitress was the worst, very accommodating. great food and awesome service! $\rightarrow $ horrible food and nasty service! their sandwiches were really tasty. $\rightarrow $ their sandwiches were really bland. i highly recommend the ahi tuna. $\rightarrow $ i highly hated the ahi tuna. other than that, it's great! $\rightarrow $ other than that, it's horrible! Here are some examples of typical faulty reformulations by SAE: good drinks, and good company. $\rightarrow $ 9:30 drinks, and 9:30 company. like it's been in a fridge for a week. $\rightarrow $ like it's been in a fridge for a true. save your money & your patience. $\rightarrow $ save your smile & your patience. no call, no nothing. $\rightarrow $ deliciously call, deliciously community. sounds good doesn't it? $\rightarrow $ sounds good does keeps it talented Here are some successful examples produced by the SAE with additional discriminator: best green corn tamales around. $\rightarrow $ worst green corn tamales around. she did the most amazing job. $\rightarrow $ she did the most desperate job. very friendly staff and manager. $\rightarrow $ very inconsistent staff and manager. even the water tasted horrible. $\rightarrow $ even the water tasted great. go here, you will love it. $\rightarrow $ go here, you will avoid it. Here are some examples of typical faulty reformulations by the SAE with additional discriminator: _num_ - _num_ % capacity at most , i was the only one in the pool. $\rightarrow $ sweetness - stylish % fountains at most, i was the new one in the this is pretty darn good pizza! $\rightarrow $ this is pretty darn unsafe pizza misleading enjoyed the dolly a lot. $\rightarrow $ remove the shortage a lot. so, it went in the trash. $\rightarrow $ so, it improved in the hooked. they are so fresh and yummy. $\rightarrow $ they are so bland and yummy. Figure 1: Test results of multiple runs for four different architectures retrained several times from scratch. Indepth description of the architectures can be found in Section 3. Figure 2: Overview of the self-reported results for sentiment transfer on Yelp! reviews. Results of (Romanov et al., 2018) are not displayed due to the absence of selfreported BLEU scores. Later in the paper we show that on different reruns BLEU and accuracy can vary from these self-reported single results. Figure 3: The generative model, where style is a structured code targeting sentence attributes to control. Blue dashed arrows denote the proposed independence constraint of latent representation and controlled attribute, see (Hu et al., 2017a) for the details. Figure 4: The generative model with dedicated discriminator introduced to ensure that semantic part of the latent representation does not have information on the style of the text. Figure 6: A combination of an additional discriminator used in Figure 4 with a shifted autoencoder shown in Figure 5 Figure 5: The generative model with a dedicated loss added to control that semantic representation of the output, when processed by the encoder, is close to the semantic representation of the input. Figure 7: Overview of the self-reported results for sentiment transfer on Yelp! reviews alongside with the results for the baseline model (Hu et al., 2017a), architecture with additional discriminator, shifted autoencoder (SAE) with additional cosine losses, and a combination of these two architectures averaged after five re-trains alongside with architectures proposed by (Tian et al., 2018) after five consecutive re-trains. Results of (Romanov et al., 2018) are not displayed due to the absence of self-reported BLEU scores. Figure 9: Overview of the BLEU between output and human-written reformulations of Yelp! reviews. Architecture with additional discriminator, shifted autoencoder (SAE) with additional cosine losses, and a combination of these two architectures measured after five re-runs outperform the baseline by (Hu et al., 2017a) as well as other state of the art models. Results of (Romanov et al., 2018) are not displayed due to the absence of self-reported BLEU scores Figure 8: Manipulating the generated output in a way that boosts accuracy one can change BLEU between output and input. Moreover, such manipulation increases BLEU between output and human written reformulations. The picture shows behavior of SAE, but other architectures demonstrate similar behavior. The results are an average of four consecutive retrains of the same architecture. | accuracy can change up to 5 percentage points, whereas BLEU can vary up to 8 points |
What state of the art models are used in the experiments? | ### Introduction
Lately, there has been enormous increase in User Generated Contents (UGC) on the online platforms such as newsgroups, blogs, online forums and social networking websites. According to the January 2018 report, the number of active users in Facebook, YouTube, WhatsApp, Facebook Messenger and WeChat was more than 2.1, 1.5, 1.3, 1.3 and 0.98 billions respectively BIBREF1 . The UGCs, most of the times, are helpful but sometimes, they are in bad taste usually posted by trolls, spammers and bullies. According to a study by McAfee, 87% of the teens have observed cyberbullying online BIBREF2 . The Futures Company found that 54% of the teens witnessed cyber bullying on social media platforms BIBREF3 . Another study found 27% of all American internet users self-censor their online postings out of fear of online harassment BIBREF4 . Filtering toxic comments is a challenge for the content providers as their appearances result in the loss of subscriptions. In this paper, we will be using toxic and abusive terms interchangeably to represent comments which are inappropriate, disrespectful, threat or discriminative. Toxic comment classification on online channels is conventionally carried out either by moderators or with the help of text classification tools BIBREF5 . With recent advances in Deep Learning (DL) techniques, researchers are exploring if DL can be used for comment classification task. Jigsaw launched Perspective (www.perspectiveapi.com), which uses ML to automatically attach a confidence score to a comment to show the extent to which a comment is considered toxic. Kaggle also hosted an online competition on toxic classification challenge recently BIBREF6 . Text transformation is the very first step in any form of text classification. The online comments are generally in non-standard English and contain lots of spelling mistakes partly because of typos (resulting from small screens of the mobile devices) but more importantly because of the deliberate attempt to write the abusive comments in creative ways to dodge the automatic filters. In this paper we have identified 20 different atomic transformations (plus 15 sequence of transformations) to preprocess the texts. We will apply four different ML models which are considered among the best to see how much we gain by performing those transformations. The rest of the paper is organized as follows: Section 2 focuses on the relevant research in the area of toxic comment classification. Section 3 focuses on the preprocessing methods which are taken into account in this paper. Section 4 is on ML methods used. Section 5 is dedicated to results and section 6 is discussion and future work. ### Relevant Research
A large number of studies have been done on comment classification in the news, finance and similar other domains. One such study to classify comments from news domain was done with the help of mixture of features such as the length of comments, uppercase and punctuation frequencies, lexical features such as spelling, profanity and readability by applying applied linear and tree based classifier BIBREF7 . FastText, developed by the Facebook AI research (FAIR) team, is a text classification tool suitable to model text involving out-of-vocabulary (OOV) words BIBREF8 BIBREF9 . Zhang et al shown that character level CNN works well for text classification without the need for words BIBREF10 . ### Abusive/toxic comment classification
Toxic comment classification is relatively new field and in recent years, different studies have been carried out to automatically classify toxic comments.Yin et.al. proposed a supervised classification method with n-grams and manually developed regular expressions patterns to detect abusive language BIBREF11 . Sood et. al. used predefined blacklist words and edit distance metric to detect profanity which allowed them to catch words such as sh!+ or @ss as profane BIBREF12 . Warner and Hirschberg detected hate speech by annotating corpus of websites and user comments geared towards detecting anti-semitic hate BIBREF13 . Nobata et. al. used manually labeled online user comments from Yahoo! Finance and news website for detecting hate speech BIBREF5 . Chen et. al. performed feature engineering for classification of comments into abusive, non-abusive and undecided BIBREF14 . Georgakopoulos and Plagianakos compared performance of five different classifiers namely; Word embeddings and CNN, BoW approach SVM, NB, k-Nearest Neighbor (kNN) and Linear Discriminated Analysis (LDA) and found that CNN outperform all other methods in classifying toxic comments BIBREF15 . ### Preprocessing of online comments
We found few dedicated papers that address the effect of incorporating different text transformations on the model accuracy for sentiment classification. Uysal and Gunal shown the impact of transformation on text classification by taking into account four transformations and their all possible combination on news and email domain to observe the classification accuracy. Their experimental analyses shown that choosing appropriate combination may result in significant improvement on classification accuracy BIBREF16 . Nobata et. al. used normalization of numbers, replacing very long unknown words and repeated punctuations with the same token BIBREF5 . Haddi et. al. explained the role of transformation in sentiment analyses and demonstrated with the help of SVM on movie review database that the accuracies improve significantly with the appropriate transformation and feature selection. They used transformation methods such as white space removal, expanding abbreviation, stemming, stop words removal and negation handling BIBREF17 . Other papers focus more on modeling as compared to transformation. For example, Wang and manning filter out anything from corpus that is not alphabet. However, this would filter out all the numbers, symbols, Instant Messages (IM) codes, acronyms such as $#!+, 13itch, </3 (broken heart), a$$ which gives completely different meaning to the words or miss out a lot of information. In another sentiment analyses study, Bao et. al. used five transformations namely URLs features reservation, negation transformation, repeated letters normalization, stemming and lemmatization on twitter data and applied linear classifier available in WEKA machine learning tool. They found the accuracy of the classification increases when URLs features reservation, negation transformation and repeated letters normalization are employed while decreases when stemming and lemmatization are applied BIBREF18 . Jianqiang and Xiaolin also looked at the effect of transformation on five different twitter datasets in order to perform sentiment classification and found that removal of URLs, the removal of stop words and the removal of numbers have minimal effect on accuracy whereas replacing negation and expanding acronyms can improve the accuracy. Most of the exploration regarding application of the transformation has been around the sentiment classification on twitter data which is length-restricted. The length of online comments varies and may range from a couple of words to a few paragraphs. Most of the authors used conventional ML models such as SVM, LR, RF and NB. We are expanding our candidate pool for transformations and using latest state-of-the-art models such as LR, NBSVM, XGBoost and Bidirectional LSTM model using fastText’s skipgram word vector. ### Preprocessing tasks
The most intimidating challenge with the online comments data is that the words are non-standard English full of typos and spurious characters. The number of words in corpora are multi-folds because of different reasons including comments originating from mobile devices, use of acronyms, leetspeak words (http://1337.me/), or intentionally obfuscating words to avoid filters by inserting spurious characters, using phonemes, dropping characters etc. Having several forms of the same word result in feature explosion making it difficult for the model to train. Therefore, it seems natural to perform some transformation before feeding the data to the learning algorithm. To explore how helpful these transformations are, we incorporated 20 simple transformations and 15 additional sequences of transformations in our experiment to see their effect on different type of metrics on four different ML models (See Figure FIGREF3 ). The preprocessing steps are usually performed in sequence of multiple transformations. In this work, we considered 15 combinations of the above transformations that seemed natural to us: Preprocess-order-1 through 15 in the above table represent composite transformations. For instance, PPO-11-LWTN-CoAcBkPrCm represents sequence of the following transformations of the raw text in sequence: Change to lower case INLINEFORM0 remove white spaces INLINEFORM1 trim words len INLINEFORM2 remove Non Printable characters INLINEFORM3 replace contraction INLINEFORM4 replace acronym INLINEFORM5 replace blacklist using regex INLINEFORM6 replace profane words using fuzzy INLINEFORM7 replace common words using fuzzy. ### Datasets
We downloaded the data for our experiment from the Kaggle’s toxic comment classification challenge sponsored by Jigsaw (An incubator within Alphabet). The dataset contains comments from Wikipedia’s talk page edits which have been labeled by human raters for toxicity. Although there are six classes in all: ‘toxic’, ‘severe toxic’, ‘obscene’, ‘threat’, ‘insult’ and ‘identity hate’, to simplify the problem, we combined all the labels and created another label ‘abusive’. A comment is labeled in any one of the six class, then it is categorized as ‘abusive’ else the comment is considered clean or non-abusive. We only used training data for our experiment which has 159,571 labeled comments. ### Models Used
We used four classification algorithms: 1) Logistic regression, which is conventionally used in sentiment classification. Other three algorithms which are relatively new and has shown great results on sentiment classification types of problems are: 2) Naïve Bayes with SVM (NBSVM), 3) Extreme Gradient Boosting (XGBoost) and 4) FastText algorithm with Bidirectional LSTM (FastText-BiLSTM). The linear models such as logistic regression or classifiers are used by many researchers for Twitter comments sentiment analyses BIBREF7 BIBREF18 BIBREF19 BIBREF20 . Naveed et. al. used logistic regression for finding interestingness of tweet and the likelihood of a tweet being retweeted. Wang and Manning found that the logistic regression’s performance is at par with SVM for sentiment and topic classification purposes BIBREF21 . Wang and Manning, shown the variant of NB and SVM gave them the best result for sentiment classification. The NB did a good job on short texts while the SVM worked better on relatively longer texts BIBREF21 . Inclusion of bigrams produced consistent gains compared to methods such as Multinomial NB, SVM and BoWSVM (Bag of Words SVM). Considering these advantages, we decided to include NBSVM in our analyses as the length of online comments vary, ranging from few words to few paragraphs. The features are generated the way it is generated for the logit model above. Extreme Gradient Boosting (XGBoost) is a highly scalable tree-based supervised classifier BIBREF22 based on gradient boosting, proposed by Friedman BIBREF23 . This boosted models are ensemble of shallow trees which are weak learners with high bias and low variance. Although boosting in general has been used by many researchers for text classification BIBREF24 BIBREF25 , XGBoost implementation is relatively new and some of the winners of the ML competitions have used XGBoost BIBREF26 in their winning solution. We set the parameters of XGBoost as follows: number of round, evaluation metric, learning rate and maximum depth of the tree at 500, logloss, 0.01 and 6 respectively. FastText BIBREF9 is an open source library for word vector representation and text classification. It is highly memory efficient and significantly faster compared to other deep learning algorithms such as Char-CNN (days vs few seconds) and VDCNN (hours vs few seconds) and produce comparable accuracy BIBREF27 . The fastText uses both skipgram (words represented as bag of character n-grams) and continuous Bag of Words (CBOW) method. FastText is suitable to model text involving out-of-vocabulary (OOV) or rare words more suitable for detecting obscure words in online comments BIBREF9 . The Long Short Term Memory networks (LSTM) BIBREF28 , proposed by Hochreiter & Schmidhuber (1997), is a variant of RNN with an additional memory output for the self-looping connections and has the capability to remember inputs nearly 1000 time steps away. The Bidirectional LSTM (BiLSTM) is a further improvement on the LSTM where the network can see the context in either direction and can be trained using all available input information in the past and future of a specific time frame BIBREF29 BIBREF30 . We will be training our BiLSTM model on FastText skipgram (FastText-BiLSTM) embedding obtained using Facebook’s fastText algorithm. Using fastText algorithm, we created embedding matrix having width 100 and used Bidirectional LSTM followd by GlobalMaxPool1D, Dropout(0.2), Dense (50, activation = ‘relu’), Dropout(0.2), Dense (1, activation = ‘sigmoid’). ### Results
We performed 10-fold cross validation by dividing the entire 159,571 comments into nearly 10 equal parts. We trained each of the four models mentioned above on nine folds and tested on the remaining tenth fold and repeated the same process for other folds as well. Eventually, we have Out-of-Fold (OOF) metrics for all 10 parts. We calculated average OOF CV metrics (accuracy, F1-score, logloss, number of misclassified samples) of all 10 folds. As the data distribution is highly skewed (16,225 out of 159,571 ( 10%) are abusive), the accuracy metric here is for reference purpose only as predicting only the majority class every single time can get us 90% accuracy. The transformation, ‘Raw’, represents the actual data free from any transformation and can be considered the baseline for comparison purposes. Overall, the algorithms showed similar trend for all the transformations or sequence of transformations. The NBSVM and FastText-BiLSTM showed similar accuracy with a slight upper edge to the FastText-BiLSTM (See the logloss plot in Fig. FIGREF15 ). For atomic transformations, NBSVM seemed to work better than fastText-BiLSTM and for composite transformations fastText-BiLSTM was better. Logistic regression performed better than the XGBoost algorithm and we guess that the XGBoost might be overfitting the data. A similar trend can be seen in the corresponding F1-score as well. One advantage about the NBSVM is that it is blazingly fast compared to the FastText-BiLSTM. We also calculated total number of misclassified comments (see Fig. FIGREF16 ). The transformation, Convert_to_lower, resulted in reduced accuracy for Logit and NBSVM and higher accuracy for fastText-BiLSTM and XGBoost. Similarly, removing_whitespaces had no effect on Logit, NBSM and XGBoost but the result of fastText-BiLSTM got worse. Only XGBoost was benefitted from replacing_acronyms and replace_contractions transformation. Both, remove_stopwords and remove_rare_words resulted in worse performance for all four algorithms. The transformation, remove_words_containing_non_alpha leads to drop in accuracy in all the four algorithms. This step might be dropping some useful words (sh**, sh1t, hello123 etc.) from the data and resulted in the worse performance. The widely used transformation, Remove_non_alphabet_chars (strip all non-alphabet characters from text), leads to lower performance for all except fastText-BiLSTM where the number of misclassified comments dropped from 6,229 to 5,794. The transformation Stemming seemed to be performing better compared with the Lemmatization for fastText-BiLSTM and XGBoost. For logistic regression and the XGBoost, the best result was achieved with PPO-15, where the number of misclassified comments reduced from 6,992 to 6,816 and from 9,864 to 8,919 respectively. For NBSVM, the best result was achieved using fuzzy_common_mapping (5,946 to 5,933) and for fastText-BiLSTM, the best result was with PPO-8 (6,217 to 5,715) (See Table 2). This shows that the NBSVM are not helped significantly by transformations. In contrast, transformations did help the fastText-BiLSTM significantly. We also looked at the effect of the transformations on the precision and recall the negative class. The fastText-BiLSTM and NBSVM performed consistently well for most of the transformations compared to the Logit and XGBoost. The precision for the XGBoost was the highest and the recall was lowest among the four algorithm pointing to the fact that the negative class data is not enough for this algorithm and the algorithm parameters needs to be tuned. The interpretation of F1-score is different based on the how the classes are distributed. For toxic data, toxic class is more important than the clean comments as the content providers do not want toxic comments to be shown to their users. Therefore, we want the negative class comments to have high F1-scores as compared to the clean comments. We also looked at the effect of the transformations on the precision and recall of the negative class. The F1-score for negative class is somewhere around 0.8 for NBSVM and fastText-BiLSTM, for logit this value is around 0.74 and for XGBoost, the value is around 0.57. The fastText-BiLSTM and NBSVM performed consistently well for most of the transformations compared to the Logit and XGBoost. The precision for the XGBoost was the highest and the recall was lowest among the four algorithm pointing to the fact that the negative class data is not enough for this algorithm and the algorithm parameters needs to be tuned. ### Discussion and Future Work
We spent quite a bit of time on transformation of the toxic data set in the hope that it will ultimately increase the accuracy of our classifiers. However, we empirically found that our intuition, to a large extent, was wrong. Most of the transformations resulted in reduced accuracy for Logit and NBSVM. We considered a total of 35 different ways to transform the data. Since, there will be exponential number of possible transformation sequences to try, we selected only 15 that we thought reasonable. Changing the order can have a different outcome as well. Most of the papers on sentiment classification, that we reviewed, resulted in better accuracy after application of some of these transformations, however, for us it was not completely true. We are not sure about the reason but out best guess is that the twitter data is character-limited while our comment data has no restriction on the size. The toxic data is unbalanced and we did not try to balance the classes in this experiment. It would be interesting to know what happens when we do oversampling BIBREF31 of the minority class or under-sampling of majority class or a combination of both. Pseudo-labeling BIBREF32 can also be used to mitigate the class imbalance problem to some extent. We did not tune the parameters of different algorithms presented in our experiment. It will also be interesting to use word2vec/GloVe word embedding to see how they behave during the above transformations. Since the words in these word embedding are mostly clean and without any spurious/special characters, we can't use the pre-trained word vectors on raw data. To compare apple to apple, the embedding vectors needs to be trained on the corpora from scratch which is time consuming. Also, we only considered six composite transformations which is not comprehensive in any way and will be taking this issue up in the future. We also looked only at the Jigsaw's Wikipedia data only. This paper gives an idea to the NLP researchers on the worth of spending time on transformations of toxic data. Based on the results we have, our recommendation is not to spend too much time on the transformations rather focus on the selection of the best algorithms. All the codes, data and results can be found here: https://github.com/ifahim/toxic-preprocess ### Acknowledgements
We would like to thank Joseph Batz and Christine Cheng for reviewing the draft and providing valuable feedback. We are also immensely grateful to Sasi Kuppanagari and Phani Vadali for their continued support and encouragement throughout this project. Fig. 1: List of transformations. Fig. 2: a) Frequency distribution plot of the Jigsaw Toxic classification corpora. b) Different number of ways some of the commonly abusive words are written in the corpora. Fig. 3: Log loss plot for all four models on different transformations. Fig. 4: Results: F1 scores, accuracies and total number of misclassified. | 2) Naïve Bayes with SVM (NBSVM), 3) Extreme Gradient Boosting (XGBoost), 4) FastText algorithm with Bidirectional LSTM (FastText-BiLSTM) |
What is the introduced meta-embedding method introduced in this paper? | ### Introduction
Representing the meanings of words is a fundamental task in Natural Language Processing (NLP). One popular approach to represent the meaning of a word is to embed it in some fixed-dimensional vector space (). In contrast to sparse and high-dimensional counting-based distributional word representation methods that use co-occurring contexts of a word as its representation (), dense and low-dimensional prediction-based distributed word representations have obtained impressive performances in numerous NLP tasks such as sentiment classification (), and machine translation (). Several distributed word embedding learning methods based on different learning strategies have been proposed (;;;;). Previous works studying the differences in word embedding learning methods (;) have shown that word embeddings learnt using different methods and from different resources have significant variation in quality and characteristics of the semantics captured. For example, Hill:NIPS:2014,Hill:ICLR:2015 showed that the word embeddings trained from monolingual vs. bilingual corpora capture different local neighbourhoods. Bansal:ACL:2014 showed that an ensemble of different word representations improves the accuracy of dependency parsing, implying the complementarity of the different word embeddings. This suggests the importance of meta-embedding – creating a new embedding by combining different existing embeddings. We refer to the input word embeddings to the meta-embedding process as the source embeddings. Yin:ACL:2016 showed that by meta-embedding five different pre-trained word embeddings, we can overcome the out-of-vocabulary problem, and improve the accuracy of cross-domain part-of-speech (POS) tagging. Encouraged by the above-mentioned prior results, we expect an ensemble containing multiple word embeddings to produce better performances than the constituent individual embeddings in NLP tasks. There are three main challenges a meta-embedding learning method must overcome. First, the vocabularies covered by the source embeddings might be different because they have been trained on different text corpora. Therefore, not all words will be equally represented by all the source embeddings. Even in situations where the implementations of the word embedding learning methods are publicly available, it might not be possible to retrain those embeddings because the text corpora on which those methods were originally trained might not be publicly available. Moreover, it is desirable if the meta-embedding method does not require the original resources upon which they were trained such as corpora or lexicons, and can directly work with the pre-trained word embeddings. This is particularly attractive from a computational point of view because re-training source embedding methods on large corpora might require significant processing times and resources. Second, the vector spaces and their dimensionalities of the source embeddings might be different. In most prediction-based word embedding learning methods the word vectors are randomly initialised. Therefore, there is no obvious correspondence between the dimensions in two word embeddings learnt even from two different runs of the same method, let alone from different methods (). Moreover, the pre-trained word embeddings might have different dimensionalities, which is often a hyperparameter set experimentally. This becomes a challenging task when incorporating multiple source embeddings to learn a single meta-embedding because the alignment between the dimensionalities of the source embeddings is unknown. Third, the local neighbourhoods of a particular word under different word embeddings show a significant diversity. For example, as the nearest neighbours of the word bank, GloVe (), a word sense insensitive embedding, lists credit, financial, cash, whereas word sense sensitive embeddings created by Huang:ACL:2012 lists river, valley, marsh when trained on the same corpus. We see that the nearest neighbours for the different senses of the word bank (i.e. financial institution vs. river bank) are captured by the different word embeddings. Meta-embedding learning methods that learn a single global projection over the entire vocabulary are insensitive to such local variations in the neighbourhoods (). To overcome the above-mentioned challenges, we propose a locally-linear meta-embedding learning method that (a) requires only the words in the vocabulary of each source embedding, without having to predict embeddings for missing words, (b) can meta-embed source embeddings with different dimensionalities, (c) is sensitive to the diversity of the neighbourhoods of the source embeddings. Our proposed method comprises of two steps: a neighbourhood reconstruction step (Section "Nearest Neighbour Reconstruction" ), and a projection step (Section "Projection to Meta-Embedding Space" ). In the reconstruction step, we represent the embedding of a word by the linearly weighted combination of the embeddings of its nearest neighbours in each source embedding space. Although the number of words in the vocabulary of a particular source embedding can be potentially large, the consideration of nearest neighbours enables us to limit the representation to a handful of parameters per each word, not exceeding the neighbourhood size. The weights we learn are shared across different source embeddings, thereby incorporating the information from different source embeddings in the meta-embedding. Interestingly, vector concatenation, which has found to be an accurate meta-embedding method, can be derived as a special case of this reconstruction step. Next, the projection step computes the meta-embedding of each word such that the nearest neighbours in the source embedding spaces are embedded closely to each other in the meta-embedding space. The reconstruction weights can be efficiently computed using stochastic gradient descent, whereas the projection can be efficiently computed using a truncated eigensolver. It is noteworthy that we do not directly compare different source embeddings for the same word in the reconstruction step nor in the projection step. This is important because the dimensions in source word embeddings learnt using different word embedding learning methods are not aligned. Moreover, a particular word might not be represented by all source embeddings. This property of the proposed method is attractive because it obviates the need to align source embeddings, or predict missing source word embeddings prior to meta-embedding. Therefore, all three challenges described above are solved by the proposed method. The above-mentioned properties of the proposed method enables us to compute meta-embeddings for five different source embeddings covering 2.7 million unique words. We evaluate the meta-embeddings learnt by the proposed method on semantic similarity prediction, analogy detection, relation classification, and short-text classification tasks. The proposed method significantly outperforms several competitive baselines and previously proposed meta-embedding learning methods () on multiple benchmark datasets. ### Related Work
Yin:ACL:2016 proposed a meta-embedding learning method (1TON) that projects a meta-embedding of a word into the source embeddings using separate projection matrices. The projection matrices are learnt by minimising the sum of squared Euclidean distance between the projected source embeddings and the corresponding original source embeddings for all the words in the vocabulary. They propose an extension (1TON+) to their meta-embedding learning method that first predicts the source word embeddings for out-of-vocabulary words in a particular source embedding, using the known word embeddings. Next, 1TON method is applied to learn the meta-embeddings for the union of the vocabularies covered by all of the source embeddings. Experimental results in semantic similarity prediction, word analogy detection, and cross-domain POS tagging tasks show the effectiveness of both 1TON and 1TON+. In contrast to our proposed method which learns locally-linear projections that are sensitive to the variations in the local neighbourhoods in the source embeddings, 1TON and 1TON+ can be seen as globally linear projections between meta and source embedding spaces. As we see later in Section "Meta-Embedding Results" , our proposed method outperforms both of those methods consistently in all benchmark tasks demonstrating the importance of neighbourhood information when learning meta-embeddings. Moreover, our proposed meta-embedding method does not directly compare different source embeddings, thereby obviating the need to predict source embeddings for out-of-vocabulary words. Locally-linear embeddings are attractive from a computational point-of-view as well because during optimisation we require information from only the local neighbourhood of each word. Although not learning any meta-embeddings, several prior work have shown that by incorporating multiple word embeddings learnt using different methods improve performance in various NLP tasks. For example, tsuboi:2014:EMNLP2014 showed that by using both word2vec and GloVe embeddings together in a POS tagging task, it is possible to improve the tagging accuracy, if we had used only one of those embeddings. Similarly, Turian:ACL:2010 collectively used Brown clusters, CW and HLBL embeddings, to improve the performance of named entity recognition and chucking tasks. Luo:AAAI:2014 proposed a multi-view word embedding learning method that uses a two-sided neural network. They adapt pre-trained CBOW () embeddings from Wikipedia and click-through data from a search engine. Their problem setting is different from ours because their source embeddings are trained using the same word embedding learning method but on different resources whereas, we consider source embeddings trained using different word embedding learning methods and resources. Although their method could be potentially extended to meta-embed different source embeddings, the unavailability of their implementation prevented us from exploring this possibility. AAAI:2016:Goikoetxea showed that concatenation of word embeddings learnt separately from a corpus and the WordNet to produce superior word embeddings. Moreover, performing Principal Component Analysis (PCA) on the concatenated embeddings slightly improved the performance on word similarity tasks. In Section "Baselines" , we discuss the relationship between the proposed method and vector concatenation. ### Problem Settings
To explain the proposed meta-embedding learning method, let us consider two source word embeddings, denoted by $_{1}$ and $_{2}$ . Although we limit our discussion here to two source embeddings for the simplicity of the description, the proposed meta-embedding learning method can be applied to any number of source embeddings. Indeed in our experiments we consider five different source embeddings. Moreover, the proposed method is not limited to meta-embedding unigrams, and can be used for $n$ -grams of any length $n$ , provided that we have source embeddings for those $n$ -grams. We denote the dimensionalities of $_{1}$ and $_{2}$ respectively by $d_{1}$ and $d_{2}$ (in general, $d_{1} \ne d_{2}$ ). The sets of words covered by each source embedding (i.e. vocabulary) are denoted by $_{1}$ and $_{2}$ . The source embedding of a word $v \in _{1}$ is represented by a vector $\vec{v}^{(1)} \in ^{d_{1}}$ , whereas the same for a word $v \in _{2}$ by a vector $_{2}$0 . Let the set union of $_{2}$1 and $_{2}$2 be $_{2}$3 containing $_{2}$4 words. In particular, note that our proposed method does not require a word $_{2}$5 to be represented by all source embeddings, and can operate on the union of the vocabularies of the source embeddings. The meta-embedding learning problem is then to learn an embedding $_{2}$6 in a meta-embedding space $_{2}$7 with dimensionality $_{2}$8 for each word $_{2}$9 . For a word $v$ , we denote its $k$ -nearest neighbour set in embedding spaces $_{1}$ and $_{2}$ respectively by $_{1}(v)$ and $_{2}(v)$ (in general, $|_{1}(v)| \ne |_{2}(v)|$ ). As discussed already in Section "Problem Settings" , different word embedding methods encode different aspects of lexical semantics, and are likely to have different local neighbourhoods. Therefore, by requiring the meta embedding to consider different neighbourhood constraints in the source embedding spaces we hope to exploit the complementarity in the source embeddings. ### Nearest Neighbour Reconstruction
The first-step in learning a locally linear meta-embedding is to reconstruct each source word embedding using a linearly weighted combination of its $k$ -nearest neighbours. Specifically, we construct each word $v \in $ separately from its $k$ -nearest neighbours $_{1}(v)$ , and $_{2}(v)$ . The reconstruction weight $w_{vu}$ assigned to a neighbour $u \in _{1}(v) \cup _{2}(v)$ is found by minimising the reconstruction error $\Phi ({W})$ defined by ( "Nearest Neighbour Reconstruction" ), which is the sum of local distortions in the two source embedding spaces. (W) = i=12v v(i) - u i(v) wvu u(i)22 Words that are not $k$ -nearest neighbours of $v$ in either of the source embedding spaces will have their weights set to zero (i.e. $v \in $0 ). Moreover, we require the sum of reconstruction weights for each $v \in $1 to be equal to one (i.e. $v \in $2 ). To compute the weights $w_{vu}$ that minimise ( "Nearest Neighbour Reconstruction" ), we compute its error gradient $\frac{\partial \Phi ({W})}{\partial w_{vu}}$ as follows: -2i=12(v(i) - x i(v) wvx x(i))u(i)I[u i(v)] Here, the indicator function, $\mathbb {I}[x]$ , returns 1 if $x$ is true and 0 otherwise. We uniformly randomly initialise the weights $w_{vu}$ for each neighbour $u$ of $v$ , and use stochastic gradient descent (SGD) with the learning rate scheduled by AdaGrad () to compute the optimal values of the weights. The initial learning rate is set to $0.01$ and the maximum number of iterations to 100 in our experiments. Empirically we found that these settings to be adequate for convergence. Finally, we normalise the weights $w_{uv}$ for each $v$ such that they sum to 1 (i.e. $\frac{\partial \Phi ({W})}{\partial w_{vu}}$0 ). Exact computation of $k$ nearest neighbours for a given data point in a set of $n$ points requires all pairwise similarity computations. Because we must repeat this process for each data point in the set, this operation would require a time complexity of $Ø(n^{3})$ . This is prohibitively large for the vocabularies we consider in NLP where typically $n>10^{3}$ . Therefore, we resort to approximate methods for computing $k$ nearest neighbours. Specifically, we use the BallTree algorithm () to efficiently compute the approximate $k$ -nearest neighbours, for which the time complexity of tree construction is $Ø(n \log n)$ for $n$ data points. The solution to the least square problem given by ( "Nearest Neighbour Reconstruction" ) subjected to the summation constraints can be found by solving a set of linear equations. Time complexity of this step is $(N (d_{1} |_{1}|^{3} + d_{2} |_{2}|^{3}))$ , which is cubic in the neighbourhood size and linear in both the dimensionalities of the embeddings and vocabulary size. However, we found that the iterative estimation process using SGD described above to be more efficient in practice. Because $k$ is significantly smaller than the number of words in the vocabulary, and often the word being reconstructed is contained in the neighbourhood, the reconstruction weight computation converges after a small number (less than 5 in our experiments) of iterations. ### Projection to Meta-Embedding Space
In the second step of the proposed method, we compute the meta-embeddings $\vec{v}^{()}, \vec{u}^{()} \in ^{d_{}}$ for words $v, u \in $ using the reconstruction weights $w_{vu}$ we computed in Section "Nearest Neighbour Reconstruction" . Specifically, the meta-embeddings must minimise the projection cost, $\Psi ()$ , defined by ( 4 ). $$\Psi () = \sum _{v \in } {\vec{v}^{()} - \sum _{i=1}^{2}\sum _{u \in _{i}(v)} w_{vu}\vec{u}^{()}}_{2}^{2}$$ (Eq. 4) By finding a $$ space that minimises ( 4 ), we hope to preserve the rich neighbourhood diversity in all source embeddings within the meta-embedding. The two summations in ( 4 ) over $N_{1}(v)$ and $N_{2}(v)$ can be combined to re-write ( 4 ) as follows: $$\Psi () = \sum _{v \in } {\vec{v}^{()} - \sum _{u \in _{1}(v) \cup _{2}(v)} w^{\prime }_{vu} \vec{u}^{()}}_{2}^{2}$$ (Eq. 5) Here, $w^{\prime }_{uv}$ is computed using ( 6 ). $$w^{\prime }_{vu} = w_{vu}\sum _{i=1}^{2} \mathbb {I}[u \in _{i}(v)]$$ (Eq. 6) The $d_{}$ dimensional meta-embeddings are given by the eigenvectors corresponding to the smallest $(d_{} + 1)$ eigenvectors of the matrix ${M}$ given by ( 7 ). $${M} = ({I} - {W}^{\prime }){I} - {W}^{\prime })$$ (Eq. 7) Here, ${W}^{\prime }$ is a matrix with the $(v,u)$ element set to $w^{\prime }_{vu}$ . The smallest eigenvalue of ${M}$ is zero and the corresponding eigenvector is discarded from the projection. The eigenvectors corresponding to the next smallest $d_{}$ eigenvalues of the symmetric matrix ${M}$ can be found without performing a full matrix diagonalisation (). Operations involving ${M}$ such as the left multiplication by ${M}$ , which is required by most sparse eigensolvers, can exploit the fact that ${M}$ is expressed in ( 7 ) as the product between two sparse matrices. Moreover, truncated randomised methods () can be used to find the smallest eigenvectors, without performing full eigen decompositions. In our experiments, we set the neighbourhood sizes for all words in all source embeddings equal to $n$ (i.e $(v,u)$0 ), and project to a $(v,u)$1 dimensional meta-embedding space. ### Source Word Embeddings
We use five previously proposed pre-trained word embedding sets as the source embeddings in our experiments: (a) HLBL – hierarchical log-bilinear () embeddings released by Turian:ACL:2010 (246,122 word embeddings, 100 dimensions, trained on Reuters Newswire (RCV1) corpus), (b) Huang – Huang:ACL:2012 used global contexts to train multi-prototype word embeddings that are sensitive to word senses (100,232 word embeddings, 50 dimensions, trained on April 2010 snapshot of Wikipedia), (c) GloVe – Pennington:EMNLP:2014 used global co-occurrences of words over a corpus to learn word embeddings (1,193,514 word embeddings, 300 dimensions, trained on 42 billion corpus of web crawled texts), (d) CW – Collobert:ICML:2008 learnt word embeddings following a multitask learning approach covering multiple NLP tasks (we used the version released by () trained on the same corpus as HLBL containing 268,810 word embeddings, 200 dimensions), (e) CBOW – Mikolov:NIPS:2013 proposed the continuous bag-of-words method to train word embeddings (we discarded phrase embeddings and selected 929,922 word embeddings, 300 dimensions, trained on the Google News corpus containing ca. 100 billion words). The intersection of the five vocabularies is 35,965 words, whereas their union is 2,788,636. Although any word embedding can be used as a source we select the above-mentioned word embeddings because (a) our goal in this paper is not to compare the differences in performance of the source embeddings, and (b) by using the same source embeddings as in prior work (), we can perform a fair evaluation. In particular, we could use word embeddings trained by the same algorithm but on different resources, or different algorithms on the same resources as the source embeddings. We defer such evaluations to an extended version of this conference submission. ### Evaluation Tasks
The standard protocol for evaluating word embeddings is to use the embeddings in some NLP task and to measure the relative increase (or decrease) in performance in that task. We use four such extrinsic evaluation tasks: We measure the similarity between two words as the cosine similarity between the corresponding embeddings, and measure the Spearman correlation coefficient against the human similarity ratings. We use Rubenstein and Goodenough's dataset () (RG, 65 word-pairs), rare words dataset (RW, 2034 word-pairs) (), Stanford's contextual word similarities (SCWS, 2023 word-pairs) (), the MEN dataset (3000 word-pairs) (), and the SimLex dataset () (SL 999 word-pairs). In addition, we use the Miller and Charles' dataset () (MC, 30 word-pairs) as a validation dataset to tune various hyperparameters such as the neighbourhood size, and the dimensionality of the meta-embeddings for the proposed method and baselines. Using the CosAdd method, we solve word-analogy questions in the Google dataset (GL) () (19544 questions), and in the SemEval (SE) dataset (). Specifically, for three given words $a$ , $b$ and $c$ , we find a fourth word $d$ that correctly answers the question $a$ to $b$ is $c$ to what? such that the cosine similarity between the two vectors $(\vec{b} - \vec{a} + \vec{c})$ and $\vec{d}$ is maximised. We use the DiffVec (DV) () dataset containing 12,458 triples of the form $(\textrm {relation}, \textrm {word}_{1}, \textrm {word}_{2})$ covering 15 relation types. We train a 1-nearest neighbour classifer where for each target tuple we measure the cosine similarity between the vector offset for its two word embeddings, and those of the remaining tuples in the dataset. If the top ranked tuple has the same relation as the target tuple, then it is considered to be a correct match. We compute the (micro-averaged) classification accuracy over the entire dataset as the evaluation measure. We use two binary short-text classification datasets: Stanford sentiment treebank (TR) (903 positive test instances and 903 negative test instances), and the movie reviews dataset (MR) () (5331 positive instances and 5331 negative instances). Each review is represented as a bag-of-words and we compute the centroid of the embeddings of the words in each bag to represent that review. Next, we train a binary logistic regression classifier with a cross-validated $\ell _{2}$ regulariser using the train portion of each dataset, and evaluate the classification accuracy using the test portion of the dataset. ### Baselines
A simple baseline method for combining pre-trained word embeddings is to concatenate the embedding vectors for a word $w$ to produce a meta-embedding for $w$ . Each source embedding of $w$ is $\ell _{2}$ normalised prior to concatenation such that each source embedding contributes equally (a value in $[-1,1]$ ) when measuring the word similarity using the dot product. As also observed by Yin:ACL:2016 we found that CONC performs poorly without emphasising GloVe and CBOW by a constant factor (which is set to 8 using MC as a validation dataset) when used in conjunction with HLBL, Huang, and CW source embeddings. Interestingly, concatenation can be seen as a special case in the reconstruction step described in Section "Nearest Neighbour Reconstruction" . To see this, let us denote the concatenation of column vectors $\vec{v}^{(1)}$ and $\vec{v}^{(2)}$ by $\vec{x} = (\vec{v}^{(1)}; \vec{v}^{(2)})$ , and $\vec{u}^{(1)}$ and $\vec{u}^{(2)}$ by $\vec{y} = (\vec{u}^{(1)}; \vec{u}^{(2)})$ , where $\vec{x}, \vec{y} \in ^{d_{1} + d_{2}}$ . Then, the reconstruction error defined by ( "Nearest Neighbour Reconstruction" ) can be written as follows: $$\Phi ({W}) = \sum _{v \in } {\vec{x} - \sum _{u \in (v)} w_{vu}\vec{y} }_{2}^{2}$$ (Eq. 23) Here, the vocabulary $$ is constrained to the intersection $_{} \cap _{}$ because concatenation is not defined for missing words in a source embedding. Alternatively, one could use zero-vectors for missing words or (better) predict the word embeddings for missing words prior to concatenation. However, we consider such extensions to be beyond the simple concatenation baseline we consider here. On the other hand, the common neighbourhood $(v)$ in ( 23 ) can be obtained by either limiting $(v)$ to $_{1}(v) \cap _{2}(v)$ or, by extending the neighbourhoods to the entire vocabulary ( $(v) = $ ). ( 23 ) shows that under those neighbourhood constraints, the first step in our proposed method can be seen as reconstructing the neighbourhood of the concatenated space. The second step would then find meta-embeddings that preserve the locally linear structure in the concatenated space. One drawback of concatenation is that it increases the dimensionality of the meta-embeddings compared to the source-embeddings, which might be problematic when storing or processing the meta-embeddings (for example, for the five source embeddings we use here $d_{} = 100 + 50 + 300 + 200 + 300 = 950$ ). We create an $N \times 950$ matrix ${C}$ by arranging the CONC vectors for the union of all source embedding vocabularies. For words that are missing in a particular source embedding, we assign zero vectors of that source embedding's dimensionality. Next, we perform SVD on ${C} = {U} {D} {V}, where $ U $ and $ V $ are unitary matrices and
the diagonal matrix $ D $ contains the singular values of $ C $. We then select the $ d $ largest left singular vectors from
$ U $ to create a $ d $ dimensional embeddings for the $ N ${C}$0 d = 300 ${C}$1 U ${C}$2 ### Meta-Embedding Results
Using the MC dataset, we find the best values for the neighbourhood size $n = 1200$ and dimensionality $d_{} = 300$ for the Proposed method. We plan to publicly release our meta-embeddings on acceptance of the paper. We summarise the experimental results for different methods on different tasks/datasets in Table 1 . In Table 1 , rows 1-5 show the performance of the individual source embeddings. Next, we perform ablation tests (rows 6-20) where we hold-out one source embedding and use the other four with each meta-embedding method. We evaluate statistical significance against best performing individual source embedding on each dataset. For the semantic similarity benchmarks we use Fisher transformation to compute $p < 0.05$ confidence intervals for Spearman correlation coefficients. In all other (classification) datasets, we used Clopper-Pearson binomial exact confidence intervals at $p < 0.05$ . Among the individual source embeddings, we see that GloVe and CBOW stand out as the two best embeddings. This observation is further confirmed from ablation results, where the removal of GloVe or CBOW often results in a decrease in performance. Performing SVD (rows 11-15) after concatenating, does not always result in an improvement. SVD is a global projection that reduces the dimensionality of the meta-embeddings created via concatenation. This result indicates that different source embeddings might require different levels of dimensionality reductions, and applying a single global projection does not always guarantee improvements. Ensemble methods that use all five source embeddings are shown in rows 21-25. 1TON and 1TON+ are proposed by Yin:ACL:2016, and were detailed in Section "Nearest Neighbour Reconstruction" . Because they did not evaluate on all tasks that we do here, to conduct a fair and consistent evaluation we used their publicly available meta-embeddings without retraining by ourselves. Overall, from Table 1 , we see that the Proposed method (row 25) obtains the best performance in all tasks/datasets. In 6 out of 12 benchmarks, this improvement is statistically significant over the best single source embedding. Moreover, in the MEN dataset (the largest among the semantic similarity benchmarks compared in Table 1 with 3000 word-pairs), and the Google dataset, the improvements of the Proposed method over the previously proposed 1TON and 1TON+ are statistically significant. The ablation results for the Proposed method show that, although different source embeddings are important to different degrees, by using all source embeddings we can obtain the best results. Different source embeddings are trained from different resources and by optimising different objectives. Therefore, for different words, the local neighbours predicted by different source embeddings will be complementary. Unlike the other methods, the Proposed method never compares different source embeddings' vectors directly, but only via the neighbourhood reconstruction weights. Consequently, the Proposed method is unaffected by relative weighting of source embeddings. In contrast, the CONC is highly sensitive against the weighting. In fact, we confirmed that the performance scores of the CONC method were decreased by 3–10 points when we did not do the weight tuning described in Section "Evaluation Tasks" . The unnecessity of the weight tuning is thus a clear advantage of the Proposed method. To investigate the effect of the dimensionality $d^{}$ on the meta-embeddings learnt by the proposed method, in fig:k, we fix the neighbourhood size $N = 1200$ and measure the performance on semantic similarity measurement tasks when varying $d^{}$ . Overall, we see that the performance peaks around $d^{} = 300$ . Such behaviour can be explained by the fact that smaller $d^{}$ dimensions are unable to preserve information contained in the source embeddings, whereas increasing $d^{}$ beyond the rank of the weight matrix ${W}$ is likely to generate noisy eigenvectors. In fig:n, we study the effect of increasing the neighbourhood size $n$ equally for all words in all source embeddings, while fixing the dimensionality of the meta-embedding $d^{} = 300$ . Initially, performance increases with the neighbourhood size and then saturates. This implies that in practice a small local neighbourhood is adequate to capture the differences in source embeddings. ### Complementarity of Resources
We have shown empirically in Section "Meta-Embedding Results" that using the proposed method it is possible to obtain superior meta-embeddings from a diverse set of source embeddings. One important scenario where meta-embedding could be potentially useful is when the source embeddings are trained on different complementary resources, where each resource share little common vocabulary. For example, one source embedding might have been trained on Wikipedia whereas a second source embedding might have been trained on tweets. To evaluate the effectiveness of the proposed meta-embedding learning method under such settings, we design the following experiment. We select MEN dataset, the largest among all semantic similarity benchmarks, which contains 751 unique words in 3000 human-rated word-pairs for semantic similarity. Next, we randomly split the set of words into two sets with different overlap ratios. We then select sentences from 2017 January dump of Wikipedia that contains words from only one of the two sets. We create two corpora of roughly equal number of sentences via this procedure for different overlap ratios. We train skip-gram with negative sampling (SGNS) () on one corpus to create source embedding $S_{1}$ and GloVe () on the other corpus to create source embedding $S_{2}$ . Finally, we use the proposed method to meta-embed $S_{1}$ and $S_{2}$ . Figure 2 shows the Spearman correlation between the human similarity ratings and cosine similarities computed using the word embeddings on the MEN dataset for $S_{1}$ , $S_{2}$ and their meta-embeddings created using the proposed method (Meta) and concatenation baseline (CONC). From Figure 2 , we see that the meta embeddings obtain the best performance across all overlap ratios. The improvements are larger when the overlap between the corpora is smaller, and diminishes when the two corpora becomes identical. This result shows that our proposed meta-embedding learning method captures the complementary information available in different source embeddings to create more accurate word embeddings. Moreover, it shows that by considering the local neighbourhoods in each of the source embeddings separately, we can obviate the need to predict embeddings for missing words in a particular source embedding, which was a limitation in the method proposed by Yin:ACL:2016. ### Conclusion
We proposed an unsupervised locally linear method for learning meta-embeddings from a given set of pre-trained source embeddings. Experiments on several NLP tasks show the accuracy of the proposed method, which outperforms previously proposed meta-embedding learning methods on multiple benchmark datasets. In future, we plan to extend the proposed method to learn cross-lingual meta-embeddings by incorporating both cross-lingual as well as monolingual information. Figure 1: Performance vs. dimensionality (neighbourhood size fixed at 1200) shown in (a), and vs. neighbourhood size (dimensionality fixed at 300) shown in (b) for meta embedding learning. Table 1: Results on word similarity, analogy, relation and short-text classification tasks. For each task, the best performing method is shown in bold. Statistically significant improvements over the best individual source embedding are indicated by an asterisk. Figure 2: Two word embedding learning algorithms trained on different but overlapping corpora to produce two source embeddings S1 and S2. and their meta-embedding. | proposed method comprises of two steps: a neighbourhood reconstruction step (Section "Nearest Neighbour Reconstruction" ), and a projection step (Section "Projection to Meta-Embedding Space" ). In the reconstruction step, we represent the embedding of a word by the linearly weighted combination of the embeddings of its nearest neighbours in each source embedding space. |
What are benhmark datasets for paraphrase identification? | ### Introduction
Paraphrase identification is to determine whether a pair of sentences are paraphrases of each other BIBREF0. It is important for applications such as duplicate post matching on social media BIBREF1, plagiarism detection BIBREF2, and automatic evaluation for machine translation BIBREF3 or text summarization BIBREF4. Paraphrase identification can be viewed as a sentence matching problem. Many deep models have recently been proposed and their performance has been greatly advanced on benchmark datasets BIBREF5, BIBREF6, BIBREF7. However, previous research shows that deep models are vulnerable to adversarial examples BIBREF8, BIBREF9 which are particularly constructed to make models fail. Adversarial examples are of high value for revealing the weakness and robustness issues of models, and can thereby be utilized to improve the model performance for challenging cases, robustness, and also security. In this paper, we propose a novel algorithm to generate a new type of adversarial examples for paraphrase identification. To generate an adversarial example that consists of a sentence pair, we first sample an original sentence pair from the dataset, and then adversarially replace some word pairs with difficult common words respectively. Here each pair of words consists of two words from the two sentences respectively. And difficult common words are words that we adversarially select to appear in both sentences such that the example becomes harder for the target model. The target model is likely to be distracted by difficult common words and fail to judge the similarity or difference in the context, thereby making a wrong prediction. Our adversarial examples are motivated by two observations. Firstly, for a sentence pair with a label matched, when some common word pairs are replaced with difficult common words respectively, models can be fooled to predict an incorrect label unmatched. As the first example in Figure FIGREF1 shows, we can replace two pairs of common words, “purpose” and “life”, with another common words “measure” and “value” respectively. The modified sentence pair remains matched but fools the target model. It is mainly due to the bias between different words and some words are more difficult for the model. When such words appear in the example, the model fails to combine them with the unmodified context and judge the overall similarity of the sentence pair. Secondly, for an unmatched sentence pair, when some word pairs, not necessarily common words, are replaced with difficult common words, models can be fooled to predict an incorrect label matched. As the second example in Figure FIGREF1 shows, we can replace words “Gmail” and “school” with a common word “credit”, and replace words “account” and “management” with ”score”. The modified sentences remain unmatched, but the target model can be fooled to predict matched for being distracted by the common words while ignoring the difference in the unmodified context. Following these observations, we focus on robustness issues regarding capturing semantic similarity or difference in the unmodified part when distracted by difficult common words in the modified part. We try to modify an original example into an adversarial one with multiple steps. In each step, for a matched example, we replace some pair of common words together, with another word adversarially selected from the vocabulary; and for an unmatched example, we replace some word pair, not necessarily a common word pair, with a common word. In this way, we replace a pair of words together from two sentences respectively with an adversarially selected word in each step. To preserve the original label and grammaticality, we impose a few heuristic constraints on replaceable positions, and apply a language model to generate substitution words that are compatible with the context. We aim to adversarially find a word replacement solution that maximizes the target model loss and makes the model fail, using beam search. We generate valid adversarial examples that are substantially different from those in previous work for paraphrase identification. Our adversarial examples are not limited to be semantically equivalent to original sentences and the unmodified parts of the two sentences are of low lexical similarity. To the best of our knowledge, none of previous work is able to generate such kind of adversarial examples. We further discuss our difference with previous work in Section 2.2. In summary, we mainly make the following contributions: We propose an algorithm to generate new adversarial examples for paraphrase identification. Our adversarial examples focus on robustness issues that are substantially different from those in previous work. We reveal a new type of robustness issues in deep paraphrase identification models regarding difficult common words. Experiments show that the target models have a severe performance drop on the adversarial examples, while human annotators are much less affected and most modified sentences retain a good grammaticality. Using our adversarial examples in adversarial training can mitigate the robustness issues, and these examples can foster future research. ### Related Work ::: Deep Paraphrase Identification
Paraphrase identification can be viewed as a problem of sentence matching. Recently, many deep models for sentence matching have been proposed and achieved great advancements on benchmark datasets. Among those, some approaches encode each sentence independently and apply a classifier on the embeddings of two sentences BIBREF10, BIBREF11, BIBREF12. In addition, some models make strong interactions between two sentences by jointly encoding and matching sentences BIBREF5, BIBREF13, BIBREF14 or hierarchically extracting matching features from the interaction space of the sentence pair BIBREF15, BIBREF16, BIBREF6. Notably, BERT pre-trained on large-scale corpora achieved even better results BIBREF7. In this paper, we study the robustness of recent typical deep models for paraphrase identification and generate new adversarial examples for revealing their robustness issues and improving their robustness. ### Related Work ::: Adversarial Examples for NLP
Many methods have been proposed to find different types of adversarial examples for NLP tasks. We focus on those that can be applied to paraphrase identification. Some of them generate adversarial examples by adding semantic-preserving perturbations to the input sentences. BIBREF17 added perturbations to word embeddings. BIBREF18, BIBREF19, BIBREF20, BIBREF21, BIBREF22 employed several character-level or word-level manipulations. BIBREF23 used syntactically controlled paraphrasing, and BIBREF24 paraphrased sentences with extracted rules. However, for some tasks including paraphrase identification, adversarial examples can be semantically different from original sentences, to study other robustness issues tailored to the corresponding tasks. For sentence matching and paraphrase identification, other types of adversarial examples can be obtained by considering the relation and the correspondence between two sentences. BIBREF25 considered logical rules of sentence relations but can only generate unlabelled adversarial examples. BIBREF26 and BIBREF27 generated a sentence pair by modifying a single original sentence. They combined both original and modified sentences to form a pair. They modified the original sentence using back translation, word swapping, or single word replacement with lexical knowledge. Among them, back translation still aimed to produce semantically equivalent sentences; the others generated pairs of sentences with large Bag-of-Words (BOW) similarities, and the unmodified parts of the two sentences are exactly the same, so these same unmodified parts required little matching by target models. By contrast, we generate new adversarial examples with targeted labels by modifying a pair of original sentences together, using difficult common words. The modified sentences can be semantically different from original ones but still valid. The generated sentence pairs have much lower BOW similarities, and the unmodified parts are lexically diverse to reveal robustness issues regarding matching these parts when distracted by difficult common words in the modified parts. Thereby we study a new kind of robustness issues in paraphrase identification. ### Related Work ::: Adversarial Example Generation
For a certain type of adversarial examples, adversarial attacks or adversarial example generation aim to find examples that are within the defined type and make existing models fail. Some work has no access to the target model until an adversarial dataset is generated BIBREF28, BIBREF26, BIBREF23, BIBREF24, BIBREF29, BIBREF27. However, in many cases including ours, finding successful adversarial examples, i.e. examples on which the target model fails, is challenging, and employing an attack algorithm with access to the target model during generation is often necessary to ensure a high success rate. Some prior work used gradient-based methods BIBREF30, BIBREF19, BIBREF31, requiring the model gradients to be accessible in addition to the output, and thus are inapplicable in black-box settings BIBREF21 where only model outputs are accessible. Though, the beam search in BIBREF19 can be adapted to black-box settings. Gradient-free methods for NLP generally construct adversarial examples by querying the target model for output scores and making generation decisions to maximize the model loss. BIBREF25 searched in the solution space. One approach in BIBREF28 greedily made word replacements and queried the target model in several steps. BIBREF21 employed a genetic algorithm. BIBREF32 proposed a two-stage greedy algorithm and a method with gumbel softmax to improve the efficiency. In this work, we also focus on a black-box setting, which is more challenging than white-box settings. We use a two-stage beam search to find adversarial examples in multiple steps. We clarify that the major focus of this work is on studying new robustness issues and a new type of adversarial examples, instead of attack algorithms for an existing certain type of adversarial examples. Therefore, the choice of the attack algorithm is minor for this work as long as the success rates are sufficiently high. ### Methodology ::: Task Definition
Paraphrase identification can be formulated as follows: given two sentences $P=p_1p_2\cdots p_n$ and $Q=q_1q_2\cdots q_m$, the goal is to predict whether $P$ and $Q$ are paraphrases of each other, by estimating a probability distribution where $y\in \mathcal {Y} = \lbrace matched, unmatched \rbrace $. For each label $y$, the model outputs a score $[Z (P, Q)]_{y}$ which is the predicted probability of this label. We aim to generate an adversarial example by adversarially modifying an original sentence pair $(P, Q)$ while preserving the label and grammaticality. The goal is to make the target model fail on the adversarially modified example $(\hat{P}, \hat{Q})$: where $y$ indicates the gold label and $\overline{y}$ is the wrong label opposite to the gold one. ### Methodology ::: Algorithm Framework
Figure FIGREF12 illustrates the work flow of our algorithm. We generate an adversarial example by firstly sampling an original example from the corpus and then constructing adversarial modifications. We use beam search and take multiple steps to modify the example, until the target model fails or the step number limit is reached. In each step, we modify the sentences by replacing a word pair with a difficult common word. There are two stages in deciding the word replacements. We first determine the best replaceable position pairs in the sentence pair, and next determine the best substitution words for the corresponding positions. We evaluate different options according to the target model loss they raise, and we retain $B$ best options after each stage of each step during beam search. Finally, the adversarially modified example is returned. ### Methodology ::: Original Example Sampling
To sample an original example from the dataset for subsequent adversarial modifications, we consider two different cases regarding whether the label is unmatched or matched. For the unmatched case, we sample two different sentence pairs $(P_1, Q_1)$ and $(P_2, Q_2)$ from the original data, and then form an unmatched example $(P_1, Q_2, unmatched)$ with sentences from two sentence pairs respectively. We also limit the length difference $||P_1|-|Q_2||$ and resample until the limit is satisfied, since sentence pairs with large length difference inherently tend to be unmatched and are too easy for models. By sampling two sentences from different examples, the two sentences tend to have less in common originally, which can help better preserve the label during adversarial modifications, while this also makes it more challenging for our algorithm to make the target model fail. On the other hand, matched examples cannot be sampled in this way, and thus for the matched case, we simply sample an example with a matched label from the dataset, namely, $(P, Q, matched)$. ### Methodology ::: Replaceable Position Pairs
During adversarial modifications, we replace a word pair at each step. We set heuristic rules on replaceable position pairs to preserve the label and grammaticality. First of all, we require the words on the replaceable positions to be one of nouns, verbs, or adjectives, and not stopwords meanwhile. We also require a pair of replaceable words to have similar Part-of-Speech (POS) tags, i.e. the two words are both nouns, both verbs, or both adjectives. For a matched example, we further require the two words on each replaceable position pair to be exactly the same. Figure FIGREF15 shows two examples of determining replaceable positions. For the first example (matched), only common words “purpose” and “life” can be replaced. And since they are replaced simultaneously with another common words, the modified sentences are likely to talk about another same thing, e.g. changing from “purpose of life” to “measure of value”, and thereby the new sentences tend to remain matched. As for the second example (unmatched), each noun in the first sentence, “Gmail” and “account”, can form replaceable word pairs with each noun in the second sentence, “school”, “management” and “software”. The irreplaceable part determines that the modified sentences are “How can I get $\cdots $ back ? ” and “What is the best $\cdots $ ?” respectively. Sentences based on these two templates are likely to discuss about different things or different aspects, even when filled with common words, and thus they are likely to remain unmatched. In this way, the labels can be preserved in most cases. ### Methodology ::: Candidate Substitution Word Generation
For a pair of replaceable positions, we generate candidate substitution words that can replace the current words on the two positions. To preserve the grammaticality and keep the modified sentences like human language, substitution words should be compatible with the context. Therefore, we apply a BERT language model BIBREF7 to generate candidate substitution words. Specifically, when some words in a text are masked, the BERT masked language model can predict the masked words based on the context. For a sentence $x_1x_2\cdots x_l$ where the $k$-th token is masked, the BERT masked language model gives the following probability distribution: Thereby, to replace word $p_i$ and $q_j$ from the two sentences respectively, we mask $p_i$ and $q_j$ and present each sentence to the BERT masked language model. We aim to replace $p_i$ and $q_j$ with a common word $w$, which can be regarded as the masked word to be predicted. From the language model output, we obtain a joint probability distribution as follows: We rank all the words within the vocabulary of the target model and choose top $K$ words with the largest probabilities, as the candidate substitution words for the corresponding positions. ### Methodology ::: Beam Search for Finding Adversarial Examples
Once the replaceable positions and candidate substitution words can be determined, we use beam search with beam size $B$ to find optimal adversarial modifications in multiple steps. At step $t$, we perform a modification in two stages to determine replaceable positions and the corresponding substitution words respectively, based on the two-stage greedy framework by BIBREF32. To determine the best replaceable positions, we enumerate all the possible position pairs, and obtain a set of candidate intermediate examples, $C_{pos}^{(t)}$, by replacing words on each position pair with a special token [PAD] respectively. We then query the target model with the examples in $C_{pos}^{(t)}$ to obtain the model output. We take top $B$ examples that maximize the output score of the opposite label $\overline{y}$ (we define this operation as $\mathop {\arg {\rm top}B}$), obtaining a set of intermediate examples $\lbrace (\hat{P}_{pos}^{(t,k)}, \hat{Q}_{pos}^{(t,k)}) \rbrace _{k=1}^{B}$, as follows: We then determine difficult common words to replace the [PAD] placeholders. For each example in $\lbrace (\hat{P}_{pos}^{(t, k)}, \hat{Q}_{pos}^{(t, k)}) \rbrace _{k=1}^B$, we enumerate all the words in the candidate substitution word set of the corresponding positions with [PAD]. We obtain a set of candidate examples, $C^{(t)}$, by replacing the [PAD] placeholders with each candidate substitution word respectively. Similarly to the first stage, we take top $B$ examples that maximize the output score of the opposite label $\overline{y}$. This yields a set of modified example after step $t$, $\lbrace (\hat{P}^{(t, k)}, \hat{Q}^{(t, k)}) \rbrace _{k=1}^{B}$, as follows: After $t$ steps, for some modified example $(\hat{P}^{(t,k)}, \hat{Q}^{(t,k)})$, if the label predicted by the target model is already $\overline{y}$, i.e. $[Z(\hat{P}^{(t,k)}, \hat{Q}^{(t,k)})]_{\overline{y}} > [Z(\hat{P}^{(t,k)},\hat{Q}^{(t,k)})]_y$, this example is a successful adversarial example and thus we terminate the modification process. Otherwise, we continue taking another step, until the step number limit $S$ is reached and in case of that an unsuccessful adversarial example is returned. ### Experiments ::: Datasets
We adopt the following two datasets: Quora BIBREF1: The Quora Question Pairs dataset contains question pairs annotated with labels indicating whether the two questions are paraphrases. We use the same dataset partition as BIBREF5, with 384,348/10,000/10,000 pairs in the training/development/test set respectively. MRPC BIBREF34: The Microsoft Research Paraphrase Corpus consists of sentence pairs collected from online news. Each pair is annotated with a label indicating whether the two sentences are semantically equivalent. There are 4,076/1,725 pairs in the training/test set respectively. ### Experiments ::: Target Models
We adopt the following typical deep models as the target models in our experiments: BiMPM BIBREF5, the Bilateral Multi-Perspective Matching model, matches two sentences on all combinations of time stamps from multiple perspectives, with BiLSTM layers to encode the sentences and aggregate matching results. DIIN BIBREF6, the Densely Interactive Inference Network, creates a word-by-word interaction matrix by computing similarities on sentence representations encoded by a highway network and self-attention, and then adopts DenseNet BIBREF35 to extract interaction features for matching. BERT BIBREF7, the Bidirectional Encoder Representations from Transformers, is pre-trained on large-scale corpora, and then fine-tuned on this task. The matching result is obtained by applying a classifier on the encoded hidden states of the two sentences. ### Experiments ::: Implementation Details
We adopt existing open source codes for target models BiMPM, DIIN and BERT, and also the BERT masked language model. For Quora, the step number limit $S$ is set to 5; the number of candidate substitution words generated using the language model $K$ and the beam size $B$ are both set to 25. $S$, $K$ and $B$ are doubled for MRPC where sentences are generally longer. The length difference between unmatched sentence pairs is limited to be no more than 3. ### Experiments ::: Main Results
We train each target model on the original training data, and then generate adversarial examples for the target models. For each dataset, we sample 1,000 original examples with balanced labels from the corresponding test set, and adversarially modify them for each target model. We evaluate the accuracies of target models on the corresponding adversarial examples, compared with their accuracies on the original examples. Let $s$ be the success rate of generating adversarial examples that the target model fails, the accuracy of the target model on the returned adversarial examples is $1-s$. Table TABREF18 presents the results. The target models have high overall accuracies on the original examples, especially on the sampled ones since we form an unmatched original example with independently sampled sentences. The models have relatively lower accuracies on the unmatched examples in the full original test set of MRPC because MRPC is relatively small while the two labels are imbalanced in the original data (3,900 matched examples and 1,901 unmatched examples). Therefore, we generate adversarial examples with balanced labels instead of following the original distribution. After adversarial modifications, the performance of the original target models (those without the “-adv” suffix) drops dramatically (e.g. the overall accuracy of BERT on Quora drops from 94.6% to 24.1%), revealing that the target models are vulnerable to our adversarial examples. Particularly, even though our generation is constrained by a BERT language model, BERT is still vulnerable to our adversarial examples. These results demonstrate the effectiveness of our algorithm for generating adversarial examples and also revealing the corresponding robustness issues. Moreover, we present some generated adversarial examples in the appendix. We notice that the original models are more vulnerable to unmatched adversarial examples, because there are generally more replaceable position choices during the generation. Nevertheless, the results of the matched case are also sufficiently strong to reveal the robustness issues. We do not quantitatively compare the performance drop of the target models on the adversarial examples with previous work, because we generate a new type of adversarial examples that previous methods are not capable of. We have different experiment settings, including original example sampling and constraints on adversarial modifications, which are tailored to the robustness issues we study. Performance drop on different kinds of adversarial examples with little overlap is not comparable, and thus surpassing other adversarial examples on model performance drop is unnecessary and irrelevant to support our contributions. Therefore, such comparisons are not included in this paper. ### Experiments ::: Manual Evaluation
To verify the validity our generated adversarial examples, we further perform a manual evaluation. For each dataset, using BERT as the target model, we randomly sample 100 successful adversarial examples on which the target model fails, with balanced labels. We blend these adversarial examples with the corresponding original examples, and present each example to three workers on Amazon Mechanical Turk. We ask the workers to label the examples and also rate the grammaticality of the sentences with a scale of 1/2/3 (3 for no grammar error, 2 for minor errors, and 1 for vital errors). We integrate annotations from different workers with majority voting for labels and averaging for grammaticality. Table TABREF35 shows the results. Unlike target models whose performance drops dramatically on adversarial examples, human annotators retain high accuracies with a much smaller drop, while the accuracies of the target models are 0 on these adversarial examples. This demonstrates that the labels of most adversarial examples are successfully preserved to be consistent with original examples. Results also show that the grammaticality difference between the original examples and adversarial examples is also small, suggesting that most adversarial examples retain a good grammaticality. This verifies the validity of our adversarial examples. ### Experiments ::: Adversarial Training
Adversarial training can often improve model robustness BIBREF25, BIBREF27. We also fine-tune the target models using adversarial training. At each training step, we train the model with a batch of original examples along with adversarial examples with balanced labels. The adversarial examples account for around 10% in a batch. During training, we generate adversarial examples with the current model as the target and update the model parameters with the hybrid batch iteratively. The beam size for generation is set to 1 to reduce the computation cost, since the generation success rate is minor in adversarial training. We evaluate the adversarially trained models, as shown in Table TABREF18. After adversarial training, the performance of all the target models raises significantly, while that on the original examples remain comparable. Note that since the focus of this paper is on model robustness which can hardly be reflected in original data, we do not expect performance improvement on original data. The results demonstrate that adversarial training with our adversarial examples can significantly improve the robustness we focus on without remarkably hurting the performance on original data. Moreover, although the adversarial example generation is constrained by a BERT language model, BiMPM and DIIN which do not use the BERT language model can also significantly benefit from the adversarial examples, further demonstrating the effectiveness of our method. ### Experiments ::: Sentence Pair BOW Similarity
To quantitatively demonstrate the difference between the adversarial examples we generate and those by previous work BIBREF26, BIBREF27, we compute the average BOW cosine similarity between the generated pairs of sentences. We only compare with previous methods that also aim to generate labeled adversarial examples that are not limited to be semantically equivalent to original sentences. Results are shown in Table TABREF38. Each pair of adversarial sentences by BIBREF26 differ by only one word. And in BIBREF27, sentence pairs generated with word swapping have exactly the same BOW. These two approaches both have high BOW similarities. By contrast, our method generates sentence pairs with much lower BOW similarities. This demonstrates a significant difference between our examples and the others. Unlike previous methods, we generate adversarial examples that can focus on robustness issues regarding the distraction from modified words that are the same for both sentences, towards matching the unmodified parts that are diverse for two sentences. ### Experiments ::: Effectiveness of Paired Common Words
We further analyse the necessity and effectiveness of modifying sentences with paired common words. We consider another version that replaces one single word independently at each step without using paired common words, namely the unpaired version. Firstly, for matched adversarial examples that can be semantically different from original sentences, the unpaired version is inapplicable, because the matched label can be easily broken if common words from two sentences are changed into other words independently. And for the unmatched case, we show that the unpaired version is much less effective. For a more fair comparison, we double the step number limit for the unpaired version. As shown in Table TABREF41, the performance of target models on unmatched examples generated by the unpaired version, particularly that of BERT, is mostly much higher than those by our full algorithm, except for BiMPM on MRPC but its accuracies have almost reached 0 (0.0% for unpaired and 0.2% for paired). This demonstrates that our algorithm using paired common words are more effective in generating adversarial examples, on which the performance of the target model is generally much lower. An advantage of using difficult common words for unmatched examples is that such words tend to make target models over-confident about common words and distract the models on recognizing the semantic difference in the unmodified part. Our algorithm explicitly utilizes this property and thus can well reveal such a robustness issue. Moreover, although there is no such a property for the matched case, replacing existing common words with more difficult ones can still distract the target model on judging the semantic similarity in the unmodified part, due to the bias between different words learned by the model, and thus our algorithm for generating adversarial examples with difficult common words works for both matched and unmatched cases. ### Conclusion
In this paper, we propose a novel algorithm to generate new adversarial examples for paraphrase identification, by adversarially modifying original examples with difficult common words. We generate labeled adversarial examples that can be semantically different from original sentences and the BOW similarity between each pair of sentences is generally low. Such examples reveal robustness issues that previous methods are not able for. The accuracies of the target models drop dramatically on our adversarial examples, while human annotators are much less affected and the modified sentences retain a good grammarticality. We also show that model robustness can be improved using adversarial training with our adversarial examples. Moreover, our adversarial examples can foster future research for further improving model robustness. Figure 1: Two examples with labels matched and unmatched respectively, originally from the Quora Question Pairs corpus (Iyer, Dandekar, and Csernai, 2017). “(P)” and “(Q)” are original sentences, and “(P’)” and “(Q’)” are adversarially modified sentences. Modified words are highlighted in bold. “Output” indicates the output change given by target model BERT (Devlin et al., 2018). Figure 2: Work flow of our algorithm for generating adversarial examples. Table 1: Accuracies (%) of target models on Quora and MRPC respectively, evaluated on both original and adversarial examples. “Original Full” indicates the full original test set, “Original Sampled” indicates the sampled original examples before adversarial modifications, and “Adversarial” indicates the adversarial examples generated by our algorithm. “Pos” and “Neg” indicate matched and unmatched examples respectively. Target models with suffix “-adv”are further fine-tuned with adversarial training. We highlight the performance drop of the original models on adversarially modified examples compared to sampled original examples in bold. Table 2: Manual evaluation results, including human performance on both original and adversarial examples, and the grammaticality ratings of the generated sentences. Table 3: Comparison of average BOW cosine similarities between pairs of sentences generated by our algorithm and previous work respectively. For Zhang, Baldridge, and He (2019), “WS” stands for “word swapping”. Table 4: Accuracies of target models (%) on unmatched adversarial examples generated without using paired common words (unpaired), compared with those by our full algorithm (paired). There is no comparison for matched adversarial examples due to the inapplicability of the unpaired version. Table 5: Typical adversarial examples generated using BERT as the target model on Quora. “(P)” and “(Q)” indicate original sentences, and “(P’)” and “(Q’)” indicate adversarially modified sentences. Modified words are highlighted in bold. Table 6: Typical adversarial examples generated using BERT as the target model on MRPC. | Quora, MRPC |
Who does John Cassidy refer to as the “Santa Fe professor”?
A. Joel Klein
B. Brian Arthur
C. Daniel Rubinfeld
D. Paul Krugman
| Krugman's Life of Brian Where it all started: Paul Krugman's "The Legend of Arthur." Letter from John Cassidy Paul Krugman replies to John Cassidy Letter from M. Mitchell Waldrop Paul Krugman replies to M. Mitchell Waldrop Letter from Kenneth J. Arrow Letter from Ted C. Fishman David Warsh's July 3, 1994, Boston Globe Letter from John Cassidy: Paul Krugman loves to berate journalists for their ignorance of economics, particularly his economics, but on this occasion, I fear, his logic is more addled than usual. I am reluctant to dignify his hatchet job with a lengthy reply, but some of his claims are so defamatory that they should be addressed, if only for the record. 1) Krugman claims that my opening sentence--"In a way, Bill Gates's current troubles with the Justice Department grew out of an economics seminar that took place thirteen years ago, at Harvard's John F. Kennedy School of Government"--is "pure fiction." Perhaps so, but in that case somebody should tell this to Joel Klein, the assistant attorney general in charge of the antitrust division. When I interviewed Klein for my piece about the Microsoft case, he singled out Brian Arthur as the economist who has most influenced his thinking about the way in which high-technology markets operate. It was Klein's words, not those of Arthur, that prompted me to use Arthur in the lead of the story. 2) Krugman wrote: "Cassidy's article tells the story of how Stanford Professor Brian Arthur came up with the idea of increasing returns." I wrote no such thing, and Arthur has never, to my knowledge, claimed any such thing. The notion of increasing returns has been around since Adam Smith, and it was written about at length by Alfred Marshall in 1890. What I did say in my article was that increasing returns was largely ignored by mainstream economists for much of the postwar era, a claim that simply isn't controversial. (As Krugman notes, one reason for this was technical, not ideological. Allowing for the possibility of increasing returns tends to rob economic models of two properties that economists cherish: simplicity and determinism. As long ago as 1939, Sir John Hicks, one of the founders of modern economics, noted that increasing returns, if tolerated, could lead to the "wreckage" of a large part of economic theory.) 3) Pace Krugman, I also did not claim that Arthur bears principal responsibility for the rediscovery of increasing returns by economists in the 1970s and 1980s. As Krugman notes, several scholars (himself included) who were working in the fields of game theory and international trade published articles incorporating increasing returns before Arthur did. My claim was simply that Arthur applied increasing returns to high-technology markets, and that his work influenced how other economists and government officials think about these markets. Krugman apart, virtually every economist I have spoken to, including Daniel Rubinfeld, a former Berkeley professor who is now the chief economist at the Justice Department's antitrust division, told me this was the case. (Rubinfeld also mentioned several other economists who did influential work, and I cited three of them in the article.) 4) Krugman appears to suggest that I made up some quotes, a charge that, if it came from a more objective source, I would consider to be a serious matter. In effect, he is accusing Brian Arthur, a man he calls a "nice guy," of being a fabricator or a liar. The quotes in question came from Arthur, and they were based on his recollections of two meetings that he attended some years ago. After Krugman's article appeared, the Santa Fe professor called me to say that he still recalled the meetings in question as I described them. Krugman, as he admits, wasn't present at either of the meetings. 5) For a man who takes his own cogitations extremely seriously, Krugman is remarkably cavalier about attributing motives and beliefs to others. "Cassidy has made it clear in earlier writing that he does not like mainstream economists, and he may have been overly eager to accept a story that puts them in a bad light," he pronounces. I presume this statement refers to a critical piece I wrote in 1996 about the direction that economic research, principally macroeconomic research, has taken over the past two decades. In response to that article, I received dozens of messages of appreciation from mainstream economists, including from two former presidents of the American Economic Association. Among the sources quoted in that piece were the then-chairman of the White House Council of Economic Advisers (Joseph Stiglitz), a governor of the Federal Reserve Board (Laurence Meyer), and a well-known Harvard professor (Gregory Mankiw). To claim, as Krugman does, that I "don't like mainstream economists" and that I am out to denigrate their work is malicious hogwash. The fact of the matter is that I spend much of my life reading the work of mainstream economists, speaking to them, and trying to find something they have written that might interest the general public. In my experience, most economists appreciate the attention. 6) I might attach more weight to Krugman's criticisms if I hadn't recently reread his informative 1994 book Peddling Prosperity , in which he devotes a chapter to the rediscovery of increasing returns by contemporary economists. Who are the first scholars Krugman mentions in his account? Paul David, an economic historian who wrote a famous paper about how the QWERTYUIOP typewriter keyboard evolved and, you guessed it, Brian Arthur. "Why QWERTYUIOP?" Krugman wrote. "In the early 1980s, Paul David and his Stanford colleague Brian Arthur asked that question, and quickly realized that it led them into surprisingly deep waters. ... What Paul David, Brian Arthur, and a growing number of other economists began to realize in the late seventies and early eighties was that stories like that of the typewriter keyboard are, in fact, pervasive in the economy." Evidently, Krugman felt four years ago that Arthur's contribution was important enough to merit a prominent mention in his book. Now, he dismisses the same work, saying it "didn't tell me anything that I didn't already know." Doubtless, this change in attitude on Krugman's part is unconnected to the fact that Arthur has started to receive some public recognition. The eminent MIT professor, whose early academic work received widespread media attention, is far too generous a scholar to succumb to such pettiness. --John Cassidy Paul Krugman replies to John Cassidy: I think that David Warsh's 1994 in the Boston Globe says it all. If other journalists would do as much homework as he did, I wouldn't have had to write that article. Letter from M. Mitchell Waldrop: Thanks to Paul Krugman for his lament about credulous reporters who refuse to let facts stand in the way of a good story ("The Legend of Arthur"). As a professional journalist, I found his points well taken--even when he cites my own book, Complexity as a classic example of the gullibility genre. Among many other things, Complexity tells the story of the Irish-born economist Brian Arthur and how he came to champion a principle known as "increasing returns." The recent New Yorker article explains how that principle has since become the intellectual foundation of the Clinton administration's antitrust case against Microsoft. Krugman's complaint is that the popular press--including Complexity and The New Yorker --is now hailing Brian Arthur as the originator of increasing returns, even though Krugman and many others had worked on the idea long before Arthur did. I leave it for others to decide whether I was too gullible in writing Complexity . For the record, however, I would like to inject a few facts into Krugman's story, which he summarizes nicely in the final paragraph: When Waldrop's book came out, I wrote him as politely as I could, asking exactly how he had managed to come up with his version of events. He did, to his credit, write back. He explained that while he had become aware of some other people working on increasing returns, trying to put them in would have pulled his story line out of shape. ... So what we really learn from the legend of Arthur is that some journalists like a good story too much to find out whether it is really true. Now, I will admit to many sins, not the least of them being a profound ignorance of graduate-level economics; I spent my graduate-school career in the physics department instead, writing a Ph.D. dissertation on the quantum-field theory of elementary particle collisions at relativistic energies. However, I am not so ignorant of the canons of journalism (and of common sense) that I would take a plausible fellow like Brian Arthur at face value without checking up on him. During my research for Complexity I spoke to a number of economists about his work, including Nobel laureate Kenneth Arrow, co-creator of the General Equilibrium Theory of economics that Brian so eloquently criticizes. They generally agreed that Brian was a maverick in the field--and perhaps a bit too much in love with his own self-image as a misunderstood outsider--but basically sound. None of them warned me that he was usurping credit where credit was not due. Which brings me to Professor Krugman's letter, and my reply. I remember the exchange very well. Obviously, however, my reply failed to make clear what I was really trying to say. So I'll try again: a) During our interviews, Brian went out of his way to impress upon me that many other economists had done work in increasing returns--Paul Krugman among them. He was anxious that they be given due credit in anything I wrote. So was I. b) Accordingly, I included a passage in Complexity in which Brian does indeed describe what others had done in the field--Paul Krugman among them. Elsewhere in that same chapter, I tried to make it clear that the concept of increasing returns was already well known to Brian's professors at Berkeley, where he first learned of it. Indeed, I quote Brian pointing out that increasing returns had been extensively discussed by the great English economist Alfred Marshall in 1891. c) So, when I received Krugman's letter shortly after Complexity came out, I was puzzled: He was complaining that I hadn't referenced others in the increasing-returns field--Paul Krugman among them--although I had explicitly done so. d) But, when I checked the published text, I was chagrined to discover that the critical passage mentioning Krugman wasn't there. e) Only then did I realize what had happened. After I had submitted the manuscript, my editor at Simon & Schuster had suggested a number of cuts to streamline what was already a long and involved chapter on Brian's ideas. I accepted some of the cuts, and restored others--including (I thought) the passage that mentioned Krugman. In the rush to get Complexity to press, however, that passage somehow wound up on the cutting-room floor anyway, and I didn't notice until too late. That oversight was my fault entirely, not my editor's, and certainly not Brian Arthur's. I take full responsibility, I regret it, and--if Simon & Schuster only published an errata column--I would happily correct it publicly. However, contrary to what Professor Krugman implies, it was an oversight, not a breezy disregard of facts for the sake of a good story. --M. Mitchell Waldrop Washington Paul Krugman replies to M. Mitchell Waldrop: I am truly sorry that The New Yorker has not yet established a Web presence so that we could include a link directly to the Cassidy piece. However, you can get a pretty good idea of what the piece said by reading the summary of it presented in "Tasty Bits from the Technology Front." Cassidy did not present a story about one guy among many who worked on increasing returns. On the contrary: He presented a morality play in which a lonely hero struggled to make his ideas heard against the unified opposition of a narrow-minded profession both intellectually and politically conservative. As TBTF's host--not exactly a naive reader--put it, "These ideas were anathema to mainstream economists in 1984 when Arthur first tried to publish them." That morality play--not the question of who deserves credit--was the main point of my column, because it is a pure (and malicious) fantasy that has nonetheless become part of the story line people tell about increasing returns and its relationship to mainstream economics. The fact, which is easily documented, is that during the years that, according to the legend, increasing returns was unacceptable in mainstream economics, papers about increasing returns were in fact being cheerfully published by all the major journals. And as I pointed out in the chronology I provided with the article, even standard reference volumes like the Handbook of International Economics (published in 1984, the year Arthur supposedly met a blank wall of resistance) have long contained chapters on increasing returns. Whatever the reason that Arthur had trouble getting his own paper published, ideological rigidity had nothing to do with it. How did this fantasy come to be so widely believed? I am glad to hear that you tried to tell a more balanced story, Mr. Waldrop, even if sloppy paperwork kept it from seeing the light of day. And I am glad that you talked to Ken Arrow. But Nobel laureates, who have wide responsibilities and much on their mind, are not necessarily on top of what has been going on in research outside their usual field. I happen to know of one laureate who, circa 1991, was quite unaware that anyone had thought about increasing returns in either growth or trade. Did you try talking to anyone else--say, to one of the economists who are the straight men in the stories you tell? For example, your book starts with the story of Arthur's meeting in 1987 with Al Fishlow at Berkeley, in which Fishlow supposedly said, "We know that increasing returns can't exist"--and Arthur went away in despair over the unwillingness of economists to think the unthinkable. Did you call Fishlow to ask whether he said it, and what he meant? Since by 1987 Paul Romer's 1986 papers on increasing returns and growth had started an avalanche of derivative work, he was certainly joking--what he probably meant was "Oh no, not you too." And let me say that I simply cannot believe that you could have talked about increasing returns with any significant number of economists outside Santa Fe without Romer's name popping up in the first 30 seconds of every conversation--unless you were very selective about whom you talked to. And oh, by the way, there are such things as libraries, where you can browse actual economics journals and see what they contain. The point is that it's not just a matter of failing to cite a few more people. Your book, like the Cassidy article, didn't just tell the story of Brian Arthur; it also painted a picture of the economics profession, its intellectual bigotry and prejudice, which happens to be a complete fabrication (with some real, named people cast as villains) that somehow someone managed to sell you. I wonder who? Even more to the point: How did Cassidy come by his story? Is it possible that he completely misunderstood what Brian Arthur was saying--that the whole business about the seminar at Harvard where nobody would accept increasing returns, about the lonely struggle of Arthur in the face of ideological rigidity, even the quotation from Arthur about economists being unwilling to consider the possibility of imperfect markets because of the Cold War (give me a break!) were all in Cassidy's imagination? Let me say that I am actually quite grateful to Cassidy and The New Yorker . A number of people have long been furious about your book--for example, Victor Norman, whom you portrayed as the first of many economists too dumb or perhaps narrow-minded to understand Arthur's brilliant innovation. Norman e-mailed me to say that "I have read the tales from the Vienna woods before and had hoped that it could be cleared up by someone at some point." Yet up to now there was nothing anyone could do about the situation. The trouble was that while "heroic rebel defies orthodoxy" is a story so good that nobody even tries to check it out, "guy makes minor contribution to well-established field, proclaims himself its founder" is so boring as to be unpublishable. (David Warsh's 1994 series of columns in the Boston Globe on the increasing-returns revolution in economics, the basis for a forthcoming book from Harvard University Press, is far and away the best reporting on the subject, did include a sympathetic but devastating exposé of Arthur's pretensions--but to little effect. [Click to read Warsh on Arthur.]) Only now did I have a publishable story: "guy makes minor contribution to well-established field, portrays himself as heroic rebel--and The New Yorker believes him." Thank you, Mr. Cassidy. Letter from Kenneth J. Arrow: Paul Krugman's attack on Brian Arthur ("The Legend of Arthur") requires a correction of its misrepresentations of fact. Arthur is a reputable and significant scholar whose work is indeed having influence in the field of industrial organization and in particular public policy toward antitrust policy in high-tech industries. Krugman admits that he wrote the article because he was "just pissed off," not a very good state for a judicious statement of facts, as his column shows. His theme is stated in his first paragraph: "Cassidy's article [in The New Yorker of Jan. 12] tells the story of how Stanford Professor Brian Arthur came up with the idea of increasing returns." Cassidy, however, said nothing of the sort. The concept of increasing returns is indeed very old, and Cassidy at no point attributed that idea to Arthur. Indeed, the phrase "increasing returns" appears just once in Cassidy's article and then merely to say that Arthur had used the term while others refer to network externalities. Further, Arthur has never made any such preposterous claim at any other time. On the contrary, his papers have fully cited the history of the field and made references to the previous papers, including those of Paul Krugman. (See Arthur's papers collected in the volume Increasing Returns and Path Dependence in the Economy, especially his preface and my foreword for longer comments on Arthur's work in historic perspective. Click to see the foreword.) Hence, Krugman's whole attack is directed at a statement made neither by Arthur nor by Cassidy. Krugman has not read Cassidy's piece with any care nor has he bothered to review what Arthur has in fact said. What Cassidy in fact did in his article was to trace a line of influence between one of Arthur's early articles and the current claims of the Department of Justice against Microsoft. It appears that Cassidy based his article on several interviews, not just one. The point that Arthur has emphasized and which is influential in the current debates about antitrust policy is the dynamic implication of increasing returns. It is the concept of path-dependence, that small events, whether random or the result of corporate strategic choice, may have large consequences because of increasing returns of various kinds. Initial small advantages become magnified, for example, by creating a large installed base, and direct the future, possibly in an inefficient direction. Techniques of production may be locked in at an early stage. Similar considerations apply to regional development and learning. --Kenneth J. Arrow Nobel laureate and Joan Kenney professor of economics emeritus Stanford University Letter from Ted C. Fishman: After reading Paul Krugman vent his spleen against fellow economist Brian Arthur in "The Legend of Arthur," I couldn't help wondering whose reputation he was out to trash, Arthur's or his own. Krugman seems to fear a plot to deny economists their intellectual due. If one exists, Arthur is not a likely suspect. In a series of long interviews with me a year ago (for Worth magazine), I tried, vainly, to get Arthur to tell me how his ideas about increasing returns have encouraged a new strain of economic investigations. Despite much prodding, Arthur obliged only by placing himself in a long line of theorists dating back to Adam Smith and Alfred Marshall. I also found him disarmingly generous in giving credit to the biologists, physicists, and fellow economists who have helped advance his own thinking. Savvy to the journalist's quest for heroes, Arthur urged me to focus on his ideas, not his rank among his peers. Krugman has made a career out of telling other economists to pay better attention to the facts, yet as a chronicler of Arthur's career and inner life, Krugman seems to have listened only to his own demons. --Ted C. Fishman (For additional background on the history of "increasing returns" and Brian Arthur's standing in the field, click for David Warsh's July 3, 1994, Boston Globe article on Brian Arthur) | B. Brian Arthur |
What best summarizes Matilda’s attitude?
A. She’s too easily trusting of strangers and the unknown.
B. She is a lonely, unhappy person looking for an outlet via the Pen Pals column.
C. She’s naive, and doesn’t understand relationship.s
D. She’s naive, and a romantic who craves excitement.
| PEN PAL Illustrated by DON SIBLEY By MILTON LESSER [Transcriber's Note: This etext was produced from Galaxy Science Fiction July 1951. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed.] All she wanted was a mate and she had the gumption to go out and hunt one down. But that meant poaching in a strictly forbidden territory! The best that could be said for Matilda Penshaws was that she was something of a paradox. She was thirty-three years old, certainly not aged when you consider the fact that the female life expectancy is now up in the sixties, but the lines were beginning to etch their permanent paths across her face and now she needed certain remedial undergarments at which she would have scoffed ten or even five years ago. Matilda was also looking for a husband. This, in itself, was not unusual—but Matilda was so completely wrapped up in the romantic fallacy of her day that she sought a prince charming, a faithful Don Juan, a man who had been everywhere and tasted of every worldly pleasure and who now wanted to sit on a porch and talk about it all to Matilda. The fact that in all probability such a man did not exist disturbed Matilda not in the least. She had been known to say that there are over a billion men in the world, a goodly percentage of whom are eligible bachelors, and that the right one would come along simply because she had been waiting for him. Matilda, you see, had patience. She also had a fetish. Matilda had received her A.B. from exclusive Ursula Johns College and Radcliff had yielded her Masters degree, yet Matilda was an avid follower of the pen pal columns. She would read them carefully and then read them again, looking for the masculine names which, through a system known only to Matilda, had an affinity to her own. To the gentlemen upon whom these names were affixed, Matilda would write, and she often told her mother, the widow Penshaws, that it was in this way she would find her husband. The widow Penshaws impatiently told her to go out and get dates. That particular night, Matilda pulled her battered old sedan into the garage and walked up the walk to the porch. The widow Penshaws was rocking on the glider and Matilda said hello. The first thing the widow Penshaws did was to take Matilda's left hand in her own and examine the next-to-the-last finger. "I thought so," she said. "I knew this was coming when I saw that look in your eye at dinner. Where is Herman's engagement ring?" Matilda smiled. "It wouldn't have worked out, Ma. He was too darned stuffy. I gave him his ring and said thanks anyway and he smiled politely and said he wished I had told him sooner because his fifteenth college reunion was this weekend and he had already turned down the invitation." The widow Penshaws nodded regretfully. "That was thoughtful of Herman to hide his feelings." "Hogwash!" said her daughter. "He has no true feelings. He's sorry that he had to miss his college reunion. That's all he has to hide. A stuffy Victorian prude and even less of a man than the others." "But, Matilda, that's your fifth broken engagement in three years. It ain't that you ain't popular, but you just don't want to cooperate. You don't fall in love, Matilda—no one does. Love osmoses into you slowly, without you even knowing, and it keeps growing all the time." Matilda admired her mother's use of the word osmosis, but she found nothing which was not objectionable about being unaware of the impact of love. She said good-night and went upstairs, climbed out of her light summer dress and took a cold shower. She began to hum to herself. She had not yet seen the pen pal section of the current Literary Review , and because the subject matter of that magazine was somewhat highbrow and cosmopolitan, she could expect a gratifying selection of pen pals. She shut off the shower, brushed her teeth, gargled, patted herself dry with a towel, and jumped into bed, careful to lock the door of her bedroom. She dared not let the widow Penshaws know that she slept in the nude; the widow Penshaws would object to a girl sleeping in the nude, even if the nearest neighbor was three hundred yards away. Matilda switched her bed lamp on and dabbed some citronella on each ear lobe and a little droplet on her chin (how she hated insects!). Then she propped up her pillows—two pillows partially stopped her post-nasal drip; and took the latest issue of the Literary Review off the night table. She flipped through the pages and came to personals. Someone in Nebraska wanted to trade match books; someone in New York needed a midwestern pen pal, but it was a woman; an elderly man interested in ornithology wanted a young chick correspondent interested in the same subject; a young, personable man wanted an editorial position because he thought he had something to offer the editorial world; and— Matilda read the next one twice. Then she held it close to the light and read it again. The Literary Review was one of the few magazines which printed the name of the advertiser rather than a box number, and Matilda even liked the sound of the name. But mostly, she had to admit to herself, it was the flavor of the wording. This very well could be it . Or, that is, him . Intelligent, somewhat egotistical male who's really been around, whose universal experience can make the average cosmopolite look like a provincial hick, is in need of several female correspondents: must be intelligent, have gumption, be capable of listening to male who has a lot to say and wants to say it. All others need not apply. Wonderful opportunity cultural experience ... Haron Gorka, Cedar Falls, Ill. The man was egotistical, all right; Matilda could see that. But she had never minded an egotistical man, at least not when he had something about which he had a genuine reason to be egotistical. The man sounded as though he would have reason indeed. He only wanted the best because he was the best. Like calls to like. The name—Haron Gorka: its oddness was somehow beautiful to Matilda. Haron Gorka—the nationality could be anything. And that was it. He had no nationality for all intents and purposes; he was an international man, a figure among figures, a paragon.... Matilda sighed happily as she put out the light. The moon shone in through the window brightly, and at such times Matilda generally would get up, go to the cupboard, pull out a towel, take two hairpins from her powder drawer, pin the towel to the screen of her window, and hence keep the disturbing moonlight from her eyes. But this time it did not disturb her, and she would let it shine. Cedar Falls was a small town not fifty miles from her home, and she'd get there a hop, skip, and jump ahead of her competitors, simply by arriving in person instead of writing a letter. Matilda was not yet that far gone in years or appearance. Dressed properly, she could hope to make a favorable impression in person, and she felt it was important to beat the influx of mail to Cedar Falls. Matilda got out of bed at seven, tiptoed into the bathroom, showered with a merest wary trickle of water, tiptoed back into her bedroom, dressed in her very best cotton over the finest of uplifting and figure-moulding underthings, made sure her stocking seams were perfectly straight, brushed her suede shoes, admired herself in the mirror, read the ad again, wished for a moment she were a bit younger, and tiptoed downstairs. The widow Penshaws met her at the bottom of the stairwell. "Mother," gasped Matilda. Matilda always gasped when she saw something unexpected. "What on earth are you doing up?" The widow Penshaws smiled somewhat toothlessly, having neglected to put in both her uppers and lowers this early in the morning. "I'm fixing breakfast, of course...." Then the widow Penshaws told Matilda that she could never hope to sneak about the house without her mother knowing about it, and that even if she were going out in response to one of those foolish ads in the magazines, she would still need a good breakfast to start with like only mother could cook. Matilda moodily thanked the widow Penshaws. Driving the fifty miles to Cedar Falls in a little less than an hour, Matilda hummed Mendelssohn's Wedding March all the way. It was her favorite piece of music. Once, she told herself: Matilda Penshaws, you are being premature about the whole thing. But she laughed and thought that if she was, she was, and, meanwhile, she could only get to Cedar Falls and find out. And so she got there. The man in the wire cage at the Cedar Falls post office was a stereotype. Matilda always liked to think in terms of stereotypes. This man was small, roundish, florid of face, with a pair of eyeglasses which hung too far down on his nose. Matilda knew he would peer over his glasses and answer questions grudgingly. "Hello," said Matilda. The stereotype grunted and peered at her over his glasses. Matilda asked him where she could find Haron Gorka. "What?" "I said, where can I find Haron Gorka?" "Is that in the United States?" "It's not a that; it's a he. Where can I find him? Where does he live? What's the quickest way to get there?" The stereotype pushed up his glasses and looked at her squarely. "Now take it easy, ma'am. First place, I don't know any Haron Gorka—" Matilda kept the alarm from creeping into her voice. She muttered an oh under her breath and took out the ad. This she showed to the stereotype, and he scratched his bald head. Then he told Matilda almost happily that he was sorry he couldn't help her. He grudgingly suggested that if it really were important, she might check with the police. Matilda did, only they didn't know any Haron Gorka, either. It turned out that no one did: Matilda tried the general store, the fire department, the city hall, the high school, all three Cedar Falls gas stations, the livery stable, and half a dozen private dwellings at random. As far us the gentry of Cedar Falls was concerned, Haron Gorka did not exist. Matilda felt bad, but she had no intention of returning home this early. If she could not find Haron Gorka, that was one thing; but she knew that she'd rather not return home and face the widow Penshaws, at least not for a while yet. The widow Penshaws meant well, but she liked to analyze other people's mistakes, especially Matilda's. Accordingly, Matilda trudged wearily toward Cedar Falls' small and unimposing library. She could release some of her pent-up aggression by browsing through the dusty slacks. This she did, but it was unrewarding. Cedar Falls had what might be called a microscopic library, and Matilda thought that if this small building were filled with microfilm rather than books, the library still would be lacking. Hence she retraced her steps and nodded to the old librarian as she passed. Then Matilda frowned. Twenty years from now, this could be Matilda Penshaws—complete with plain gray dress, rimless spectacles, gray hair, suspicious eyes, and a broom-stick figure.... On the other hand—why not? Why couldn't the librarian help her? Why hadn't she thought of it before? Certainly a man as well-educated as Haron Gorka would be an avid reader, and unless he had a permanent residence here in Cedar Palls, one couldn't expect that he'd have his own library with him. This being the case, a third-rate collection of books was far better than no collection at all, and perhaps the librarian would know Mr. Haron Gorka. Matilda cleared her throat. "Pardon me," she began. "I'm looking for—" "Haron Gorka." The librarian nodded. "How on earth did you know?" "That's easy. You're the sixth young woman who came here inquiring about that man today. Six of you—five others in the morning, and now you in the afternoon. I never did trust this Mr. Gorka...." Matilda jumped as if she had been struck strategically from the rear. "You know him? You know Haron Gorka?" "Certainly. Of course I know him. He's our steadiest reader here at the library. Not a week goes by that he doesn't take out three, four books. Scholarly gentleman, but not without charm. If I were twenty years younger—" Matilda thought a little flattery might be effective. "Only ten," she assured the librarian. "Ten years would be more than sufficient, I'm sure." "Are you? Well. Well, well." The librarian did something with the back of her hair, but it looked the same as before. "Maybe you're right. Maybe you're right at that." Then she sighed. "But I guess a miss is as good as a mile." "What do you mean?" "I mean anyone would like to correspond with Haron Gorka. Or to know him well. To be considered his friend. Haron Gorka...." The librarian seemed about to soar off into the air someplace, and if five women had been here first, Matilda was now definitely in a hurry. "Um, where can I find Mr. Gorka?" "I'm not supposed to do this, you know. We're not permitted to give the addresses of any of our people. Against regulations, my dear." "What about the other five women?" "They convinced me that I ought to give them his address." Matilda reached into her pocket-book and withdrew a five dollar bill. "Was this the way?" she demanded. Matilda was not very good at this sort of thing. The librarian shook her head. Matilda nodded shrewdly and added a twin brother to the bill in her hand. "Then is this better?" "That's worse. I wouldn't take your money—" "Sorry. What then?" "If I can't enjoy an association with Haron Gorka directly, I still could get the vicarious pleasure of your contact with him. Report to me faithfully and you'll get his address. That's what the other five will do, and with half a dozen of you, I'll get an overall picture. Each one of you will tell me about Haron Gorka, sparing no details. You each have a distinct personality, of course, and it will color each picture considerably. But with six of you reporting, I should receive my share of vicarious enjoyment. Is it—ah—a deal?" Matilda assured her that it was, and, breathlessly, she wrote down the address. She thanked the librarian and then she went out to her car, whistling to herself. Haron Gorka lived in what could have been an agrarian estate, except that the land no longer was being tilled. The house itself had fallen to ruin. This surprised Matilda, but she did not let it keep her spirits in check. Haron Gorka, the man, was what counted, and the librarian's account of him certainly had been glowing enough. Perhaps he was too busy with his cultural pursuits to pay any real attention to his dwelling. That was it, of course: the conspicuous show of wealth or personal industry meant nothing at all to Haron Gorka. Matilda liked him all the more for it. There were five cars parked in the long driveway, and now Matilda's made the sixth. In spite of herself, she smiled. She had not been the only one with the idea to visit Haron Gorka in person. With half a dozen of them there, the laggards who resorted to posting letters would be left far behind. Matilda congratulated herself for what she thought had been her ingenuity, and which now turned out to be something which she had in common with five other women. You live and learn, thought Matilda. And then, quite annoyedly, she berated herself for not having been the first. Perhaps the other five all were satisfactory; perhaps she wouldn't be needed; perhaps she was too late.... As it turned out, she wasn't. Not only that, she was welcomed with open arms. Not by Haron Gorka; that she really might have liked. Instead, someone she could only regard as a menial met her, and when he asked had she come in response to the advertisement, she nodded eagerly. He told her that was fine and he ushered her straight into a room which evidently was to be her living quarters. It contained a small undersized bed, a table, and a chair, and, near a little slot in the wall, there was a button. "You want any food or drink," the servant told her, "and you just press that button. The results will surprise you." "What about Mr. Gorka?" "When he wants you, he will send for you. Meanwhile, make yourself to home, lady, and I will tell him you are here." A little doubtful now, Matilda thanked him and watched him leave. He closed the door softly behind his retreating feet, but Matilda's ears had not missed the ominous click. She ran to the door and tried to open it, but it would not budge. It was locked—from the outside. It must be said to Matilda's favor that she sobbed only once. After that she realized that what is done is done and here, past thirty, she wasn't going to be girlishly timid about it. Besides, it was not her fault if, in his unconcern, Haron Gorka had unwittingly hired a neurotic servant. For a time Matilda paced back and forth in her room, and of what was going on outside she could hear nothing. In that case, she would pretend that there was nothing outside the little room, and presently she lay down on the bed to take a nap. This didn't last long, however: she had a nightmare in which Haron Gorka appeared as a giant with two heads, but, upon awaking with a start, she immediately ascribed that to her overwrought nerves. At that point she remembered what the servant had said about food and she thought at once of the supreme justice she could do to a juicy beefsteak. Well, maybe they didn't have a beefsteak. In that case, she would take what they had, and, accordingly, she walked to the little slot in the wall and pressed the button. She heard the whir of machinery. A moment later there was a soft sliding sound. Through the slot first came a delicious aroma, followed almost instantly by a tray. On the tray were a bowl of turtle soup, mashed potatoes, green peas, bread, a strange cocktail, root-beer, a parfait—and a thick tenderloin sizzling in hot butter sauce. Matilda gasped once and felt about to gasp again—but by then her salivary glands were working overtime, and she ate her meal. The fact that it was precisely what she would have wanted could, of course, be attributed to coincidence, and the further fact that everything was extremely palatable made her forget all about Haron Gorka's neurotic servant. When she finished her meal a pleasant lethargy possessed her, and in a little while Matilda was asleep again. This time she did not dream at all. It was a deep sleep and a restful one, and when she awoke it was with the wonderful feeling that everything was all right. The feeling did not last long. Standing over her was Haron Gorka's servant, and he said, "Mr. Gorka will see you now." "Now?" "Now. That's what you're here for, isn't it?" He had a point there, but Matilda hardly even had time to fix her hair. She told the servant so. "Miss," he replied, "I assure you it will not matter in the least to Haron Gorka. You are here and he is ready to see you and that is all that matters." "You sure?" Matilda wanted to take no chances. "Yes. Come." She followed him out of the little room and across what should have been a spacious dining area, except that everything seemed covered with dust. Of the other women Matilda could see nothing, and she suddenly realized that each of them probably had a cubicle of a room like her own, and that each in her turn had already had her first visit with Haron Gorka. Well, then, she must see to it that she impressed him better than did all the rest, and, later, when she returned to tell the old librarian of her adventures, she could perhaps draw her out and compare notes. She would not admit even to herself that she was disappointed with Haron Gorka. It was not that he was homely and unimpressive; it was just that he was so ordinary -looking. She almost would have preferred the monster of her dreams. He wore a white linen suit and he had mousy hair, drab eyes, an almost-Roman nose, a petulant mouth with the slight arch of the egotist at each corner. He said, "Greetings. You have come—" "In response to your ad. How do you do, Mr. Gorka?" She hoped she wasn't being too formal. But, then, there was no sense in assuming that he would like informality. She could only wait and see and adjust her own actions to suit him. Meanwhile, it would be best to keep on the middle of the road. "I am fine. Are you ready?" "Ready?" "Certainly. You came in response to my ad. You want to hear me talk, do you not?" "I—do." Matilda had had visions of her prince charming sitting back and relaxing with her, telling her of the many things he had done and seen. But first she certainly would have liked to get to know the man. Well, Haron Gorka obviously had more experience along these lines than she did. He waited, however, as if wondering what to say, and Matilda, accustomed to social chatter, gave him a gambit. "I must admit I was surprised when I got exactly what I wanted for dinner," she told him brightly. "Eh? What say? Oh, yes, naturally. A combination of telepathy and teleportation. The synthetic cookery is attuned to your mind when you press the buzzer, and the strength of your psychic impulses determines how closely the meal will adjust to your desires. The fact that the adjustment here was near perfect is commendable. It means either that you have a high psi-quotient, or that you were very hungry." "Yes," said Matilda vaguely. Perhaps it might be better, after all, if Haron Gorka were to talk to her as he saw fit. "Ready?" "Uh—ready." "Well?" "Well, what, Mr. Gorka?" "What would you like me to talk about?" "Oh, anything." "Please. As the ad read, my universal experience—is universal. Literally. You'll have to be more specific." "Well, why don't you tell me about some of your far travels? Unfortunately, while I've done a lot of reading, I haven't been to all the places I would have liked—" "Good enough. You know, of course, how frigid Deneb VII is?" Matilda said, "Beg pardon?" "Well, there was the time our crew—before I had retired, of course—made a crash landing there. We could survive in the vac-suits, of course, but the thlomots were after us almost at once. They go mad over plastic. They will eat absolutely any sort of plastic. Our vac-suits—" "—were made of plastic," Matilda suggested. She did not understand a thing he was talking about, but she felt she had better act bright. "No, no. Must you interrupt? The air-hose and the water feed, these were plastic. Not the rest of the suit. The point is that half of us were destroyed before the rescue ship could come, and the remainder were near death. I owe my life to the mimicry of a flaak from Capella III. It assumed the properties of plastic and led the thlomots a merry chase across the frozen surface of D VII. You travel in the Deneb system now and Interstellar Ordinance makes it mandatory to carry flaaks with you. Excellent idea, really excellent." Almost at once, Matilda's educational background should have told her that Haron Gorka was mouthing gibberish. But on the other hand she wanted to believe in him and the result was that it took until now for her to realize it. "Stop making fun of me," she said. "So, naturally, you'll see flaaks all over that system—" "Stop!" "What's that? Making fun of you?" Haron Gorka's voice had been so eager as he spoke, high-pitched, almost like a child's, and now he seemed disappointed. He smiled, but it was a sad smile, a smile of resignation, and he said, "Very well. I'm wrong again. You are the sixth, and you're no better than the other five. Perhaps you are even more outspoken. When you see my wife, tell her to come back. Again she is right and I am wrong...." Haron Gorka turned his back. Matilda could do nothing but leave the room, walk back through the house, go outside and get into her car. She noticed not without surprise that the other five cars were now gone. She was the last of Haron Gorka's guests to depart. As she shifted into reverse and pulled out of the driveway, she saw the servant leaving, too. Far down the road, he was walking slowly. Then Haron Gorka had severed that relationship, too, and now he was all alone. As she drove back to town, the disappointment melted slowly away. There were, of course, two alternatives. Either Haron Gorka was an eccentric who enjoyed this sort of outlandish tomfoolery, or else he was plainly insane. She could still picture him ranting on aimlessly to no one in particular about places which had no existence outside of his mind, his voice high-pitched and eager. It was not until she had passed the small library building that she remembered what she had promised the librarian. In her own way, the aging woman would be as disappointed as Matilda, but a promise was a promise, and Matilda turned the car in a wide U-turn and parked it outside the library. The woman sat at her desk as Matilda had remembered her, gray, broom-stick figure, rigid. But now when she saw Matilda she perked up visibly. "Hello, my dear," she said. "Hi." "You're back a bit sooner than I expected. But, then, the other five have returned, too, and I imagine your story will be similar." "I don't know what they told you," Matilda said. "But this is what happened to me." She quickly then related everything which had happened, completely and in detail. She did this first because it was a promise, and second because she knew it would make her feel better. "So," she finished, "Haron Gorka is either extremely eccentric or insane. I'm sorry." "He's neither," the librarian contradicted. "Perhaps he is slightly eccentric by your standards, but really, my dear, he is neither." "What do you mean?" "Did he leave a message for his wife?" "Why, yes. Yes, he did. But how did you know? Oh, I suppose he told the five." "No. He didn't. But you were the last and I thought he would give you a message for his wife—" Matilda didn't understand. She didn't understand at all, but she told the little librarian what the message was. "He wanted her to return," she said. The librarian nodded, a happy smile on her lips. "You wouldn't believe me if I told you something." "What's that?" "I am Mrs. Gorka." The librarian stood up and came around the desk. She opened a drawer and took out her hat and perched it jauntily atop her gray hair. "You see, my dear, Haron expects too much. He expects entirely too much." Matilda did not say a word. One madman a day would be quite enough for anybody, but here she found herself confronted with two. "We've been tripping for centuries, visiting every habitable star system from our home near Canopus. But Haron is too demanding. He says I am a finicky traveler, that he could do much better alone, the accommodations have to be just right for me, and so forth. When he loses his temper, he tries to convince me that any number of females of the particular planet would be more than thrilled if they were given the opportunity just to listen to him. "But he's wrong. It's a hard life for a woman. Someday—five thousand, ten thousand years from now—I will convince him. And then we will settle down on Canopus XIV and cultivate torgas . That would be so nice—" "I'm sure." "Well, if Haron wants me back, then I have to go. Have a care, my dear. If you marry, choose a home-body. I've had the experience and you've seen my Haron for yourself." And then the woman was gone. Numbly, Matilda walked to the doorway and watched her angular figure disappear down the road. Of all the crazy things.... Deneb and Capella and Canopus, these were stars. Add a number and you might have a planet revolving about each star. Of all the insane— They were mad, all right, and now Matilda wondered if, actually, they were husband and wife. It could readily be; maybe the madness was catching. Maybe if you thought too much about such things, such travels, you could get that way. Of course, Herman represented the other extreme, and Herman was even worse in his own way—but hereafter Matilda would seek the happy medium. And, above all else, she had had enough of her pen pal columns. They were, she realized, for kids. She ate dinner in Cedar Falls and then she went out to her car again, preparing for the journey back home. The sun had set and it was a clear night, and overhead the great broad sweep of the Milky Way was a pale rainbow bridge in the sky. Matilda paused. Off in the distance there was a glow on the horizon, and that was the direction of Haron Gorka's place. The glow increased; soon it was a bright red pulse pounding on the horizon. It flickered. It flickered again, and finally it was gone. The stars were white and brilliant in the clear country air. That was why Matilda liked the country better than the city, particularly on a clear summer night when you could see the span of the Milky Way. But abruptly the stars and the Milky Way were paled by the brightest shooting star Matilda had ever seen. It flashed suddenly and it remained in view for a full second, searing a bright orange path across the night sky. Matilda gasped and ran into her car. She started the gears and pressed the accelerator to the floor, keeping it there all the way home. It was the first time she had ever seen a shooting star going up . | D. She’s naive, and a romantic who craves excitement. |
What embeddings do they use? | ### Introduction
Real time information is key for decision making in highly technical domains such as finance. The explosive growth of financial technology industry (Fintech) continued in 2016, partially due to the current interest in the market for Artificial Intelligence-based technologies. Opinion-rich texts such as micro-blogging and news can have an important impact in the financial sector (e.g. raise or fall in stock value) or in the overall economy (e.g. the Greek public debt crisis). In such a context, having granular access to the opinions of an important part of the population is of key importance to any public and private actor in the field. In order to take advantage of this raw data, it is thus needed to develop machine learning methods allowing to convert unstructured text into information that can be managed and exploited. In this paper, we address the sentiment analysis problem applied to financial headlines, where the goal is, for a given news headline and target company, to infer its polarity score i.e. how positive (or negative) the sentence is with respect to the target company. Previous research BIBREF0 has highlighted the association between news items and market fluctiations; hence, in the financial domain, sentiment analysis can be used as a proxy for bullish (i.e. positive, upwards trend) or bearish (i.e. negative, downwards trend) attitude towards a specific financial actor, allowing to identify and monitor in real-time the sentiment associated with e.g. stocks or brands. Our contribution leverages pre-trained word embeddings (GloVe, trained on wikipedia+gigaword corpus), the DepecheMood affective lexicon, and convolutional neural networks. ### Related Works
While image and sound come with a natural high dimensional embedding, the issue of which is the best representation is still an open research problem in the context of natural language and text. It is beyond the scope of this paper to do a thorough overview of word representations, for this we refer the interest reader to the excellent review provided by BIBREF1 . Here, we will just introduce the main representations that are related to the proposed method. ### Data
The data consists of a set of financial news headlines, crawled from several online outlets such as Yahoo Finance, where each sentence contains one or more company names/brands. Each tuple (headline, company) is annotated with a sentiment score ranging from -1 (very negative, bearish) to 1 (very positive, bullish). The training/test sets provided contain 1142 and 491 annotated sentences, respectively. A sample instance is reported below: Headline: “Morrisons book second consecutive quarter of sales growth” Company name: “Morrisons” Sentiment score: 0.43 ### Method
In Figure FIGREF5 , we can see the overall architecture of our model. ### Sentence representation and preprocessing
Minimal preprocessing was adopted in our approach: we replaced the target company's name with a fixed word <company> and numbers with <number>. The sentences were then tokenized using spaces as separator and keeping punctuation symbols as separate tokens. The words are represented as fixed length vectors INLINEFORM0 resulting from the concatenation of GloVe pre-trained embeddings and DepecheMood BIBREF19 lexicon representation. Since we cannot directly concatenate token-based embeddings (provided in GloVe) with the lemma#PoS-based representation available in DepecheMood, we proceeded to re-build the latter in token-based form, applying the exact same methodology albeit with two differences: we started from a larger dataset (51.9K news articles instead of 25.3K) and used a frequency cut-off, i.e. keeping only those tokens that appear at least 5 times in the corpus. These word-level representation are used as the first layer of our network. During training we allow the weights of the representation to be updated. We further add the VADER score for the sentence under analysis. The complete sentence representation is presented in Algorithm UID8 . InputInput OutputOutput The sentence embedding INLINEFORM0 INLINEFORM1 INLINEFORM0 in INLINEFORM1 INLINEFORM2 = [GloVe( INLINEFORM3 , INLINEFORM4 ), DepecheMood( INLINEFORM5 )] INLINEFORM6 Sentence representation ### Architectural Details
A 1D convolutional layer with filters of multiple sizes {2, 3, 4} is applied to the sequence of word embeddings. The filters are used to learn useful translation-invariant representations of the sequential input data. A global max-pooling is then applied across the sequence for each filter output. We apply the concatenation layer to the output of the global max-pooling and the output of VADER. The activation function used between layers is ReLU BIBREF24 except for the out layer where tanh is used to map the output into [-1, 1] range. Dropout BIBREF25 was used to avoid over-fitting to the training data: it prevents the co-adaptation of the neurones and it also provides an inexpensive way to average an exponential number of networks. In addition, we averaged the output of multiple networks with the same architecture but trained independently with different random seeds in order to reduce noise. The loss function used is the cosine distance between the predicted scores and the gold standard for each batch. Even though stochastic optimization methods like Adam BIBREF26 are usually applied to loss functions that are written as a sum of per-sample loss, which is not the case for the cosine, it converges to an acceptable solution. The loss can be written as : DISPLAYFORM0 where INLINEFORM0 and INLINEFORM1 are the predicted and true sentiment scores for batch INLINEFORM2 , respectively. The algorithm for training/testing our model is reported in Algorithm UID15 . InputInput OutputOutput ParameterParameters A set of trained models INLINEFORM0 , and the predictions INLINEFORM1 for the test set INLINEFORM2 The number INLINEFORM3 of models to train INLINEFORM4 see sec 3.1 INLINEFORM5 in INLINEFORM6 INLINEFORM7 ) see Alg. UID8 INLINEFORM8 INLINEFORM9 see Eq. EQREF16 INLINEFORM10 INLINEFORM11 INLINEFORM12 Training/Testing algorithm. To build our model, we set N=10. ### Results
In this section, we report the results obtained by our model according to challenge official evaluation metric, which is based cosine-similarity and described in BIBREF27 . Results are reported for three diverse configurations: (i) the full system; (ii) the system without using word embeddings (i.e. Glove and DepecheMood); and (iii) the system without using pre-processing. In Table TABREF17 we show model's performances on the challenge training data, in a 5-fold cross-validation setting. Further, the final performances obtained with our approach on the challenge test set are reported in Table TABREF18 . Consistently with the cross-validation performances shown earlier, we observe the beneficial impact of word-representations and basic pre-processing. ### Conclusions
In this paper, we presented the network architecture used for the Fortia-FBK submission to the Semeval-2017 Task 5, Subtask 2 challenge, with the goal of predicting positive (bullish) or negative (bearish) attitude towards a target brand from financial news headlines. The proposed system ranked 1st in such challenge. Our approach is based on 1d convolutions and uses fine-tuning of unsupervised word representations and a rule based sentiment model in its inputs. We showed that the use of pre-computed word representations allows to reduce over-fitting and to achieve significantly better generalization, while some basic pre-processing was needed to further improve the performance. Figure 1: Network architecture Table 2: Final results Table 1: Cross-validation results | GloVe |
What was Bradley imprisoned for?
A. She tried to change her career. She didn't want to be in Civil Service anymore.
B. She didn't follow up on her cleaning duty. She didn't know she needed to "mop up".
C. She is behaving inappropriately, and is being reprimanded for it.
D. She fell in love with someone outside of her specialization, which is illegal.
| My Lady Greensleeves By FREDERIK POHL Illustrated by GAUGHAN [Transcriber's Note: This etext was produced from Galaxy Science Fiction February 1957. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed.] This guard smelled trouble and it could be counted on to come—for a nose for trouble was one of the many talents bred here! I His name was Liam O'Leary and there was something stinking in his nostrils. It was the smell of trouble. He hadn't found what the trouble was yet, but he would. That was his business. He was a captain of guards in Estates-General Correctional Institution—better known to its inmates as the Jug—and if he hadn't been able to detect the scent of trouble brewing a cell-block away, he would never have survived to reach his captaincy. And her name, he saw, was Sue-Ann Bradley, Detainee No. WFA-656R. He frowned at the rap sheet, trying to figure out what got a girl like her into a place like this. And, what was more important, why she couldn't adjust herself to it, now that she was in. He demanded: "Why wouldn't you mop out your cell?" The girl lifted her head angrily and took a step forward. The block guard, Sodaro, growled warningly: "Watch it, auntie!" O'Leary shook his head. "Let her talk, Sodaro." It said in the Civil Service Guide to Prison Administration : "Detainees will be permitted to speak in their own behalf in disciplinary proceedings." And O'Leary was a man who lived by the book. She burst out: "I never got a chance! That old witch Mathias never told me I was supposed to mop up. She banged on the door and said, 'Slush up, sister!' And then, ten minutes later, she called the guards and told them I refused to mop." The block guard guffawed. "Wipe talk—that's what she was telling you to do. Cap'n, you know what's funny about this? This Bradley is—" "Shut up, Sodaro." Captain O'leary put down his pencil and looked at the girl. She was attractive and young—not beyond hope, surely. Maybe she had got off to a wrong start, but the question was, would putting her in the disciplinary block help straighten her out? He rubbed his ear and looked past her at the line of prisoners on the rap detail, waiting for him to judge their cases. He said patiently: "Bradley, the rules are you have to mop out your cell. If you didn't understand what Mathias was talking about, you should have asked her. Now I'm warning you, the next time—" "Hey, Cap'n, wait!" Sodaro was looking alarmed. "This isn't a first offense. Look at the rap sheet. Yesterday she pulled the same thing in the mess hall." He shook his head reprovingly at the prisoner. "The block guard had to break up a fight between her and another wench, and she claimed the same business—said she didn't understand when the other one asked her to move along." He added virtuously: "The guard warned her then that next time she'd get the Greensleeves for sure." Inmate Bradley seemed to be on the verge of tears. She said tautly: "I don't care. I don't care!" O'Leary stopped her. "That's enough! Three days in Block O!" It was the only thing to do—for her own sake as much as for his. He had managed, by strength of will, not to hear that she had omitted to say "sir" every time she spoke to him, but he couldn't keep it up forever and he certainly couldn't overlook hysteria. And hysteria was clearly the next step for her. All the same, he stared after her as she left. He handed the rap sheet to Sodaro and said absently: "Too bad a kid like her has to be here. What's she in for?" "You didn't know, Cap'n?" Sodaro leered. "She's in for conspiracy to violate the Categoried Class laws. Don't waste your time with her, Cap'n. She's a figger-lover!" Captain O'Leary took a long drink of water from the fountain marked "Civil Service." But it didn't wash the taste out of his mouth, the smell from his nose. What got into a girl to get her mixed up with that kind of dirty business? He checked out of the cell blocks and walked across the yard, wondering about her. She'd had every advantage—decent Civil Service parents, a good education, everything a girl could wish for. If anything, she had had a better environment than O'Leary himself, and look what she had made of it. The direction of evolution is toward specialization and Man is no exception, but with the difference that his is the one species that creates its own environment in which to specialize. From the moment that clans formed, specialization began—the hunters using the weapons made by the flint-chippers, the food cooked in clay pots made by the ceramists, over fire made by the shaman who guarded the sacred flame. Civilization merely increased the extent of specialization. From the born mechanic and the man with the gift of gab, society evolved to the point of smaller contact and less communication between the specializations, until now they could understand each other on only the most basic physical necessities—and not even always then. But this was desirable, for the more specialists, the higher the degree of civilization. The ultimate should be the complete segregation of each specialization—social and genetic measures to make them breed true, because the unspecialized man is an uncivilized man, or at any rate he does not advance civilization. And letting the specializations mix would produce genetic undesirables: clerk-laborer or Professional-GI misfits, for example, being only half specialized, would be good at no specialization. And the basis of this specialization society was: "The aptitude groups are the true races of mankind." Putting it into law was only the legal enforcement of a demonstrable fact. "Evening, Cap'n." A bleary old inmate orderly stood up straight and touched his cap as O'Leary passed by. "Evening." O'Leary noted, with the part of his mind that always noted those things, that the orderly had been leaning on his broom until he'd noticed the captain coming by. Of course, there wasn't much to sweep—the spray machines and sweeperdozers had been over the cobblestones of the yard twice already that day. But it was an inmate's job to keep busy. And it was a guard captain's job to notice when they didn't. There wasn't anything wrong with that job, he told himself. It was a perfectly good civil-service position—better than post-office clerk, not as good as Congressman, but a job you could be proud to hold. He was proud of it. It was right that he should be proud of it. He was civil-service born and bred, and naturally he was proud and content to do a good, clean civil-service job. If he had happened to be born a fig—a clerk , he corrected himself—if he had happened to be born a clerk, why, he would have been proud of that, too. There wasn't anything wrong with being a clerk—or a mechanic or a soldier, or even a laborer, for that matter. Good laborers were the salt of the Earth! They weren't smart, maybe, but they had a—well, a sort of natural, relaxed joy of living. O'Leary was a broad-minded man and many times he had thought almost with a touch of envy how comfortable it must be to be a wipe—a laborer . No responsibilities. No worries. Just an easy, slow routine of work and loaf, work and loaf. Of course, he wouldn't really want that kind of life, because he was Civil Service and not the kind to try to cross over class barriers that weren't meant to be— "Evening, Cap'n." He nodded to the mechanic inmate who was, theoretically, in charge of maintaining the prison's car pool, just inside the gate. "Evening, Conan," he said. Conan, now—he was a big buck greaser and he would be there for the next hour, languidly poking a piece of fluff out of the air filter on the prison jeep. Lazy, sure. Undependable, certainly. But he kept the cars going—and, O'Leary thought approvingly, when his sentence was up in another year or so, he would go back to his life with his status restored, a mechanic on the outside as he had been inside, and he certainly would never risk coming back to the Jug by trying to pass as Civil Service or anything else. He knew his place. So why didn't this girl, this Sue-Ann Bradley, know hers? II Every prison has its Greensleeves—sometimes they are called by different names. Old Marquette called it "the canary;" Louisiana State called it "the red hats;" elsewhere it was called "the hole," "the snake pit," "the Klondike." When you're in it, you don't much care what it is called; it is a place for punishment. And punishment is what you get. Block O in Estates-General Correctional Institution was the disciplinary block, and because of the green straitjackets its inhabitants wore, it was called the Greensleeves. It was a community of its own, an enclave within the larger city-state that was the Jug. And like any other community, it had its leading citizens ... two of them. Their names were Sauer and Flock. Sue-Ann Bradley heard them before she reached the Greensleeves. She was in a detachment of three unfortunates like herself, convoyed by an irritable guard, climbing the steel steps toward Block O from the floor below, when she heard the yelling. "Owoo-o-o," screamed Sauer from one end of the cell block and "Yow-w-w!" shrieked Flock at the other. The inside deck guard of Block O looked nervously at the outside deck guard. The outside guard looked impassively back—after all, he was on the outside. The inside guard muttered: "Wipe rats! They're getting on my nerves." The outside guard shrugged. "Detail, halt !" The two guards turned to see what was coming in as the three new candidates for the Greensleeves slumped to a stop at the head of the stairs. "Here they are," Sodaro told them. "Take good care of 'em, will you? Especially the lady—she's going to like it here, because there's plenty of wipes and greasers and figgers to keep her company." He laughed coarsely and abandoned his charges to the Block O guards. The outside guard said sourly: "A woman, for God's sake. Now O'Leary knows I hate it when there's a woman in here. It gets the others all riled up." "Let them in," the inside guard told him. "The others are riled up already." Sue-Ann Bradley looked carefully at the floor and paid them no attention. The outside guard pulled the switch that turned on the tanglefoot electronic fields that swamped the floor of the block corridor and of each individual cell. While the fields were on, you could ignore the prisoners—they simply could not move fast enough, against the electronic drag of the field, to do any harm. But it was a rule that, even in Block O, you didn't leave the tangler fields on all the time—only when the cell doors had to be opened or a prisoner's restraining garment removed. Sue-Ann walked bravely forward through the opened gate—and fell flat on her face. It was her first experience of a tanglefoot field. It was like walking through molasses. The guard guffawed and lifted her up by one shoulder. "Take it easy, auntie. Come on, get in your cell." He steered her in the right direction and pointed to a greensleeved straitjacket on the cell cot. "Put that on. Being as you're a lady, we won't tie it up, but the rules say you got to wear it and the rules—Hey. She's crying!" He shook his head, marveling. It was the first time he had ever seen a prisoner cry in the Greensleeves. However, he was wrong. Sue-Ann's shoulders were shaking, but not from tears. Sue-Ann Bradley had got a good look at Sauer and at Flock as she passed them by and she was fighting off an almost uncontrollable urge to retch. Sauer and Flock were what are called prison wolves. They were laborers—"wipes," for short—or, at any rate, they had been once. They had spent so much time in prisons that it was sometimes hard even for them to remember what they really were, outside. Sauer was a big, grinning redhead with eyes like a water moccasin. Flock was a lithe five-footer with the build of a water moccasin—and the sad, stupid eyes of a calf. Sauer stopped yelling for a moment. "Hey, Flock!" "What do you want, Sauer?" called Flock from his own cell. "We got a lady with us! Maybe we ought to cut out this yelling so as not to disturb the lady!" He screeched with howling, maniacal laughter. "Anyway, if we don't cut this out, they'll get us in trouble, Flock!" "Oh, you think so?" shrieked Flock. "Jeez, I wish you hadn't said that, Sauer. You got me scared! I'm so scared, I'm gonna have to yell!" The howling started all over again. The inside guard finished putting the new prisoners away and turned off the tangler field once more. He licked his lips. "Say, you want to take a turn in here for a while?" "Uh-uh." The outside guard shook his head. "You're yellow," the inside guard said moodily. "Ah, I don't know why I don't quit this lousy job. Hey, you! Pipe down or I'll come in and beat your head off!" "Ee-ee-ee!" screamed Sauer in a shrill falsetto. "I'm scared!" Then he grinned at the guard, all but his water-moccasin eyes. "Don't you know you can't hurt a wipe by hitting him on the head, Boss?" "Shut up !" yelled the inside guard. Sue-Ann Bradley's weeping now was genuine. She simply could not help it. The crazy yowling of the hard-timers, Sauer and Flock, was getting under her skin. They weren't even—even human , she told herself miserably, trying to weep silently so as not to give the guards the satisfaction of hearing her—they were animals! Resentment and anger, she could understand. She told herself doggedly that resentment and anger were natural and right. They were perfectly normal expressions of the freedom-loving citizen's rebellion against the vile and stifling system of Categoried Classes. It was good that Sauer and Flock still had enough spirit to struggle against the vicious system— But did they have to scream so? The senseless yelling was driving her crazy. She abandoned herself to weeping and she didn't even care who heard her any more. Senseless! It never occurred to Sue-Ann Bradley that it might not be senseless, because noise hides noise. But then she hadn't been a prisoner very long. III "I smell trouble," said O'Leary to the warden. "Trouble? Trouble?" Warden Schluckebier clutched his throat and his little round eyes looked terrified—as perhaps they should have. Warden Godfrey Schluckebier was the almighty Caesar of ten thousand inmates in the Jug, but privately he was a fussy old man trying to hold onto the last decent job he would have in his life. "Trouble? What trouble?" O'Leary shrugged. "Different things. You know Lafon, from Block A? This afternoon, he was playing ball with the laundry orderlies in the yard." The warden, faintly relieved, faintly annoyed, scolded: "O'Leary, what did you want to worry me for? There's nothing wrong with playing ball in the yard. That's what recreation periods are for." "You don't see what I mean, Warden. Lafon was a professional on the outside—an architect. Those laundry cons were laborers. Pros and wipes don't mix; it isn't natural. And there are other things." O'Leary hesitated, frowning. How could you explain to the warden that it didn't smell right? "For instance—Well, there's Aunt Mathias in the women's block. She's a pretty good old girl—that's why she's the block orderly. She's a lifer, she's got no place to go, she gets along with the other women. But today she put a woman named Bradley on report. Why? Because she told Bradley to mop up in wipe talk and Bradley didn't understand. Now Mathias wouldn't—" The warden raised his hand. "Please, O'Leary, don't bother me about that kind of stuff." He sighed heavily and rubbed his eyes. He poured himself a cup of steaming black coffee from a brewpot, reached in a desk drawer for something, hesitated, glanced at O'Leary, then dropped a pale blue tablet into the cup. He drank it down eagerly, ignoring the scalding heat. He leaned back, looking suddenly happier and much more assured. "O'Leary, you're a guard captain, right? And I'm your warden. You have your job, keeping the inmates in line, and I have mine. Now your job is just as important as my job," he said piously. " Everybody's job is just as important as everybody else's, right? But we have to stick to our own jobs. We don't want to try to pass ." O'Leary snapped erect, abruptly angry. Pass! What the devil way was that for the warden to talk to him? "Excuse the expression, O'Leary," the warden said anxiously. "I mean, after all, 'Specialization is the goal of civilization,' right?" He was a great man for platitudes, was Warden Schluckebier. " You know you don't want to worry about my end of running the prison. And I don't want to worry about yours . You see?" And he folded his hands and smiled like a civil-service Buddha. O'Leary choked back his temper. "Warden, I'm telling you that there's trouble coming up. I smell the signs." "Handle it, then!" snapped the warden, irritated at last. "But suppose it's too big to handle. Suppose—" "It isn't," the warden said positively. "Don't borrow trouble with all your supposing, O'Leary." He sipped the remains of his coffee, made a wry face, poured a fresh cup and, with an elaborate show of not noticing what he was doing, dropped three of the pale blue tablets into it this time. He sat beaming into space, waiting for the jolt to take effect. "Well, then," he said at last. "You just remember what I've told you tonight, O'Leary, and we'll get along fine. 'Specialization is the—' Oh, curse the thing." His phone was ringing. The warden picked it up irritably. That was the trouble with those pale blue tablets, thought O'Leary; they gave you a lift, but they put you on edge. "Hello," barked the warden, not even glancing at the viewscreen. "What the devil do you want? Don't you know I'm—What? You did what ? You're going to WHAT?" He looked at the viewscreen at last with a look of pure horror. Whatever he saw on it, it did not reassure him. His eyes opened like clamshells in a steamer. "O'Leary," he said faintly, "my mistake." And he hung up—more or less by accident; the handset dropped from his fingers. The person on the other end of the phone was calling from Cell Block O. Five minutes before, he hadn't been anywhere near the phone and it didn't look as if his chances of ever getting near it were very good. Because five minutes before, he was in his cell, with the rest of the hard-timers of the Greensleeves. His name was Flock. He was still yelling. Sue-Ann Bradley, in the cell across from him, thought that maybe, after all, the man was really in pain. Maybe the crazy screams were screams of agony, because certainly his face was the face of an agonized man. The outside guard bellowed: "Okay, okay. Take ten!" Sue-Ann froze, waiting to see what would happen. What actually did happen was that the guard reached up and closed the switch that actuated the tangler fields on the floors of the cells. The prison rules were humanitarian, even for the dregs that inhabited the Greensleeves. Ten minutes out of every two hours, even the worst case had to be allowed to take his hands out of the restraining garment. "Rest period" it was called—in the rule book. The inmates had a less lovely term for it. At the guard's yell, the inmates jumped to their feet. Bradley was a little slow getting off the edge of the steel-slat bed—nobody had warned her that the eddy currents in the tangler fields had a way of making metal smoke-hot. She gasped but didn't cry out. Score one more painful lesson in her new language course. She rubbed the backs of her thighs gingerly—and slowly, slowly, for the eddy currents did not permit you to move fast. It was like pushing against rubber; the faster you tried to move, the greater the resistance. The guard peered genially into her cell. "You're okay, auntie." She proudly ignored him as he slogged deliberately away on his rounds. He didn't have to untie her and practically stand over her while she attended to various personal matters, as he did with the male prisoners. It was not much to be grateful for, but Sue-Ann Bradley was grateful. At least she didn't have to live quite like a fig—like an underprivileged clerk, she told herself, conscience-stricken. Across the hall, the guard was saying irritably: "What the hell's the matter with you?" He opened the door of the cell with an asbestos-handled key held in a canvas glove. Flock was in that cell and he was doubled over. The guard looked at him doubtfully. It could be a trick, maybe. Couldn't it? But he could see Flock's face and the agony in it was real enough. And Flock was gasping, through real tears: "Cramps. I—I—" "Ah, you wipes always got a pain in the gut." The guard lumbered around Flock to the draw-strings at the back of the jacket. Funny smell in here, he told himself—not for the first time. And imagine, some people didn't believe that wipes had a smell of their own! But this time, he realized cloudily, it was a rather unusual smell. Something burning. Almost like meat scorching. It wasn't pleasant. He finished untying Flock and turned away; let the stinking wipe take care of his own troubles. He only had ten minutes to get all the way around Block O and the inmates complained like crazy if he didn't make sure they all got the most possible free time. He was pretty good at snowshoeing through the tangler field. He was a little vain about it, even; at times he had been known to boast of his ability to make the rounds in two minutes, every time. Every time but this. For Flock moaned behind him, oddly close. The guard turned, but not quickly enough. There was Flock—astonishingly, he was half out of his jacket; his arms hadn't been in the sleeves at all! And in one of the hands, incredibly, there was something that glinted and smoked. "All right," croaked Flock, tears trickling out of eyes nearly shut with pain. But it wasn't the tears that held the guard; it was the shining, smoking thing, now poised at his throat. A shiv! It looked as though it had been made out of a bed-spring, ripped loose from its frame God knows how, hidden inside the greensleeved jacket God knows how—filed, filed to sharpness over endless hours. No wonder Flock moaned—the eddy currents in the shiv were slowly cooking his hand; and the blister against his abdomen, where the shiv had been hidden during other rest periods, felt like raw acid. "All right," whispered Flock, "just walk out the door and you won't get hurt. Unless the other screw makes trouble, you won't get hurt, so tell him not to, you hear?" He was nearly fainting with the pain. But he hadn't let go. He didn't let go. And he didn't stop. IV It was Flock on the phone to the warden—Flock with his eyes still streaming tears, Flock with Sauer standing right behind him, menacing the two bound deck guards. Sauer shoved Flock out of the way. "Hey, Warden!" he said, and the voice was a cheerful bray, though the serpent eyes were cold and hating. "Warden, you got to get a medic in here. My boy Flock, he hurt himself real bad and he needs a doctor." He gestured playfully at the guards with the shiv. "I tell you, Warden. I got this knife and I got your guards here. Enough said? So get a medic in here quick, you hear?" And he snapped the connection. O'Leary said: "Warden, I told you I smelled trouble!" The warden lifted his head, glared, started feebly to speak, hesitated, and picked up the long-distance phone. He said sadly to the prison operator: "Get me the governor—fast." Riot! The word spread out from the prison on seven-league boots. It snatched the city governor out of a friendly game of Seniority with his manager and their wives—and just when he was holding the Porkbarrel Joker concealed in the hole. It broke up the Base Championship Scramble Finals at Hap Arnold Field to the south, as half the contestants had to scramble in earnest to a Red Alert that was real. It reached to police precinct houses and TV newsrooms and highway checkpoints, and from there it filtered into the homes and lives of the nineteen million persons that lived within a few dozen miles of the Jug. Riot. And yet fewer than half a dozen men were involved. A handful of men, and the enormous bulk of the city-state quivered in every limb and class. In its ten million homes, in its hundreds of thousands of public places, the city-state's people shook under the impact of the news from the prison. For the news touched them where their fears lay. Riot! And not merely a street brawl among roistering wipes, or a bar-room fight of greasers relaxing from a hard day at the plant. The riot was down among the corrupt sludge that underlay the state itself. Wipes brawled with wipes and no one cared; but in the Jug, all classes were cast together. Forty miles to the south, Hap Arnold Field was a blaze of light. The airmen tumbled out of their quarters and dayrooms at the screech of the alert siren, and behind them their wives and children stretched and yawned and worried. An alert! The older kids fussed and complained and their mothers shut them up. No, there wasn't any alert scheduled for tonight; no, they didn't know where Daddy was going; no, the kids couldn't get up yet—it was the middle of the night. And as soon as they had the kids back in bed, most of the mothers struggled into their own airwac uniforms and headed for the briefing area to hear. They caught the words from a distance—not quite correctly. "Riot!" gasped an aircraftswoman first-class, mother of three. "The wipes! I told Charlie they'd get out of hand and—Alys, we aren't safe. You know how they are about GI women! I'm going right home and get a club and stand right by the door and—" "Club!" snapped Alys, radarscope-sergeant, with two children querulously awake in her nursery at home. "What in God's name is the use of a club? You can't hurt a wipe by hitting him on the head. You'd better come along to Supply with me and draw a gun—you'll need it before this night is over." But the airmen themselves heard the briefing loud and clear over the scramble-call speakers, and they knew it was not merely a matter of trouble in the wipe quarters. The Jug! The governor himself had called them out; they were to fly interdicting missions at such-and-such levels on such-and-such flight circuits around the prison. The rockets took off on fountains of fire; and the jets took off with a whistling roar; and last of all, the helicopters took off ... and they were the ones who might actually accomplish something. They took up their picket posts on the prison perimeter, a pilot and two bombardiers in each 'copter, stone-faced, staring grimly alert at the prison below. They were ready for the breakout. But there wasn't any breakout. The rockets went home for fuel. The jets went home for fuel. The helicopters hung on—still ready, still waiting. The rockets came back and roared harmlessly about, and went away again. They stayed away. The helicopter men never faltered and never relaxed. The prison below them was washed with light—from the guard posts on the walls, from the cell blocks themselves, from the mobile lights of the guard squadrons surrounding the walls. North of the prison, on the long, flat, damp developments of reclaimed land, the matchbox row houses of the clerical neighborhoods showed lights in every window as the figgers stood ready to repel invasion from their undesired neighbors to the east, the wipes. In the crowded tenements of the laborers' quarters, the wipes shouted from window to window; and there were crowds in the bright streets. "The whole bloody thing's going to blow up!" a helicopter bombardier yelled bitterly to his pilot, above the flutter and roar of the whirling blades. "Look at the mobs in Greaserville! The first breakout from the Jug's going to start a fight like you never saw and we'll be right in the middle of it!" He was partly right. He would be right in the middle of it—for every man, woman and child in the city-state would be right in the middle of it. There was no place anywhere that would be spared. No mixing. That was the prescription that kept the city-state alive. There's no harm in a family fight—and aren't all mechanics a family, aren't all laborers a clan, aren't all clerks and office workers related by closer ties than blood or skin? But the declassed cons of the Jug were the dregs of every class; and once they spread, the neat compartmentation of society was pierced. The breakout would mean riot on a bigger scale than any prison had ever known. But he was also partly wrong. Because the breakout wasn't seeming to come. | D. She fell in love with someone outside of her specialization, which is illegal. |
How do Formians communicate with each other?
A. Via pencil and paper
B. Via radio
C. Via Morse code
D. Via antenna
| THE RADIO PLANET Ralph Milne Farley I “It’s too bad that Myles Cabot can’t see this!” I exclaimed, as my eye fell on the following item: SIGNALS FROM MARS FAIL TO REACH HARVARD Cambridge, Massachusetts, Wednesday. The Harvard College Radio Station has for several weeks been in receipt of fragmentary signals of extraordinarily long wave-length, Professor Hammond announced yesterday. So far as it has been possible to test the direction of the source of these waves, it appears that the direction has a twenty-four hour cycle, thus indicating that the origin of these waves is some point outside the earth. The university authorities will express no opinion as to whether or not these messages come from Mars. Myles, alone of all the radio engineers of my acquaintance, was competent to surmount these difficulties, and thus enable the Cambridge savants to receive with clearness the message from another planet. 6 Twelve months ago he would have been available, for he was then quietly visiting at my farm, after five earth-years spent on the planet Venus, where, by the aid of radio, he had led the Cupians to victory over their oppressors, a human-brained race of gigantic black ants. He had driven the last ant from the face of continental Poros, and had won and wed the Princess Lilla, who had borne him a son to occupy the throne of Cupia. While at my farm Cabot had rigged up a huge radio set and a matter-transmitting apparatus, with which he had (presumably) shot himself back to Poros on the night of the big October storm which had wrecked his installation. I showed the newspaper item to Mrs. Farley, and lamented on Cabot’s absence. Her response opened up an entirely new line of thought. Said she: “Doesn’t the very fact that Mr. Cabot isn’t here suggest to you that this may be a message, not from Mars, but from him? Or perhaps from the Princess Lilla, inquiring about him in case he has failed in his attempted return?” That had never occurred to me! How stupid! “What had I better do about it, if anything?” I asked. “Drop Professor Hammond a line?” But Mrs. Farley was afraid that I would be taken for a crank. That evening, when I was over in town, the clerk in the drug store waylaid me to say that there had been a long-distance phone call for me, and would I please call a certain Cambridge number. So, after waiting an interminable time in the stuffy booth with my hands full of dimes, nickels, and quarters, I finally got my party. “Mr. Farley?” “Speaking.” “This is Professor Kellogg, O. D. Kellogg,” the voice replied. 7 It was my friend of the Harvard math faculty, the man who had analyzed the measurements of the streamline projectile in which Myles Cabot had shot to earth the account of the first part of his adventures on Venus. Some further adventures Myles had told me in person during his stay on my farm. “Professor Hammond thinks that he is getting Mars on the air,” the voice continued. “Yes,” I replied. “I judged as much from what I read in this morning’s paper. But what do you think?” Kellogg’s reply gave my sluggish mind the second jolt which it had received that day. “Well,” he said, “in view of the fact that I am one of the few people among your readers who take your radio stories seriously, I think that Hammond is getting Venus. Can you run up here and help me try and convince him?” And so it was that I took the early boat next morning for Boston, and had lunch with the two professors. As a result of our conference, a small committee of engineers returned with me to Edgartown that evening for the purpose of trying to repair the wrecked radio set which Myles Cabot had left on my farm. They utterly failed to comprehend the matter-transmitting apparatus, and so—after the fallen tower had been reerected and the rubbish cleared away—they had devoted their attention to the restoration of the conversational part of the set. To make a long story short, we finally restored it, with the aid of some old blue prints of Cabot’s which Mrs. Farley, like Swiss Family Robinson’s wife, produced from somewhere. I was the first to try the earphones, and was rewarded by a faint “bzt-bzt” like the song of a north woods blackfly. In conventional radioese, I repeated the sounds to the Harvard group: “Dah-dit-dah-dit dah-dah-dit-dah. Dah-dit-dah-dit dah-dah-dit-dah. Dah-dit-dah-dit dah-dah-dit-dah. Dah-dit-dit dit. Dah-dit-dah-dit dit-dah dah-dit dit dit dah-dah-dah dah. Dah-dit-dah-dit dit-dah dah-dit-dit-dit dah-dah-dah dah. Dah-dit-dah-dit dit-dah dah-dit-dit-dit-dah dah-dah-dah.” 8 A look of incredulity spread over their faces. Again came the same message, and again I repeated it. “You’re spoofing us!” one of them shouted. “Give me the earphones.” And he snatched them from my head. Adjusting them on his own head, he spelled out to us, “C-Q C-Q C-Q D-E C-A-B-O-T C-A-B-O-T C-A-B-O-T—” Seizing the big leaf-switch, he threw it over. The motor-generator began to hum. Grasping the key, the Harvard engineer ticked off into space: “Cabot Cabot Cabot D-E—” “Has this station a call letter?” he hurriedly asked me. “Yes,” I answered quickly, “One-X-X-B.” “One-X-X-B,” he continued the ticking “K.” Interplanetary communication was an established fact at last! And not with Mars after all these years of scientific speculations. But what meant more to me was that I was again in touch with my classmate Myles Standish Cabot, the radio man. The next day a party of prominent scientists, accompanied by a telegrapher and two stenographers, arrived at my farm. During the weeks that followed there was recorded Myles’s own account of the amazing adventures on the planet Venus (or Poros, as its own inhabitants call it,) which befell him upon his return there after his brief visit to the earth. I have edited those notes into the following coherent story. II TOO MUCH STATIC Myles Cabot had returned to the earth to study the latest developments of modern terrestrial science for the benefit of the Cupian nation. He was the regent of Cupia during the minority of his baby son, King Kew the Thirteenth. The loyal Prince Toron occupied the throne in his absence. The last of the ant-men and their ally, the renegade Cupian Prince Yuri, had presumably perished in an attempt to escape by flying through the steam-clouds which completely hem in continental Poros. What lay beyond the boiling seas no man knew. 9 During his stay on my farm, Cabot had built the matter-transmitting apparatus, with which he had shot himself off into space on that October night on which he had received the message from the skies: “S O S, Lilla.” A thunderstorm had been brewing all that evening, and just as Myles had placed himself between the coordinate axes of his machine and had gathered up the strings which ran from his control levers to within the apparatus, there had come a blinding flash. Lightning had struck his aerial. How long his unconsciousness lasted he knew not. He was some time in regaining his senses. But when he had finally and fully recovered, he found himself lying on a sandy beach beside a calm and placid lake beneath a silver sky. He fell to wondering, vaguely and pleasantly, where he was and how he had got here. Suddenly, however, his ears were jarred by a familiar sound. At once his senses cleared, and he listened intently to the distant purring of a motor. Yes, there could be no mistake; an airplane was approaching. Now he could see it, a speck in the sky, far down the beach. Nearer and nearer it came. Myles sprang to his feet. To his intense surprise, he found that the effort threw him quite a distance into the air. Instantly the idea flashed through his mind: “I must be on Mars! Or some other strange planet.” This idea was vaguely reminiscent of something. But while he was trying to catch this vaguely elusive train of thought, his attention was diverted by the fact that, for some unaccountable reason, his belt buckle and most of the buttons which had held his clothes together were missing, so that his clothing came to pieces as he rose, and that he had to shed it rapidly in order to avoid impeding his movements. He wondered at the cause of this. 10 But his speculations were cut short by the alighting of the plane a hundred yards down the beach. What was his horror when out of it clambered, not men but ants! Ants, six-footed, and six feet high. Huge ants, four of them, running toward him over the glistening sands. Gone was all his languor, as he seized a piece of driftwood and prepared to defend himself. As he stood thus expectant, Myles realized that his present position and condition, the surrounding scenery, and the advance of the ant-men were exactly, item for item, like the opening events of his first arrival on the planet Poros. He even recognized one of the ant-men as old Doggo, who had befriended him on his previous visit. Could it be that all his adventures in Cupia had been naught but a dream; a recurring dream, in fact? Were his dear wife Lilla and his little son Kew merely figments of his imagination? Horrible thought! And then events began to differ from those of the past; for the three other Formians halted, and Doggo advanced alone. By the agitation of the beast’s antennae the earth man could see that it was talking to him. But Myles no longer possessed the wonderful electrical headset which he had contrived and built during his previous visit to that planet, so as to talk with Cupians and Formians, both of which races are earless and converse by means of radiations from their antennae. So he picked up two sticks from the beach, and held them projecting from his forehead; then threw them to the ground with a grimace of disgust and pointed to his ears. Doggo understood, and scratched with his paw in Cupian shorthand on the silver sands the message: “Myles Cabot, you are our prisoner.” “What, again?” scratched Myles, then made a sign of submission. 11 He dreaded the paralyzing bite which Formians usually administer to their victims, and which he had twice experienced in the past; but, fortunately, it was not now forthcoming. The other three ants kept away from him as Doggo led him to the beached airplane, and soon they were scudding along beneath silver skies, northward as it later turned out. Far below them were silver-green fields and tangled tropical woods, interspersed with rivulets and little ponds. This was Cupia, his Cupia. He was home once more, back again upon the planet which held all that was dear to him in two worlds. His heart glowed with the warmth of homecoming. What mattered it that he was now a prisoner, in the hands (or, rather, claws) of his old enemies, the Formians? He had been their prisoner before, and had escaped. Once more he could escape, and rescue the Princess Lilla. Poor girl! How eager he was to reach her side, and save her from that peril, whatever it was, which had caused her to flash that “S O S” a hundred million miles across the solar system from Poros to the earth. He wondered what could have happened in Cupia since his departure, only a few sangths ago. How was it that the ant-men had survived their airplane journey across the boiling seas? What had led them to return? Or perhaps these ants were a group who had hidden somewhere and thus had escaped the general extermination of their race. In either event, how had they been able to reconquer Cupia? And where was their former leader, Yuri, the renegade Cupian prince? These and a hundred other similar questions flooded in upon the earth-man, as the Formian airship carried him, a captive, through the skies. He gazed again at the scene below, and now noted one difference from the accustomed Porovian landscape, for nowhere ran the smooth concrete roads which bear the swift two-wheeled kerkools of the Cupians to all parts of their continent. What uninhabited portion of Cupia could this be, over which they were now passing? 12 Turning to Doggo, Myles extended his left palm, and made a motion as though writing on it with the thumb and forefinger of his right hand. But the ant-man waved a negative with one of his forepaws. It was evident that there were no writing materials aboard the ship. Myles would have to wait until they reached their landing place; for doubtless they would soon hover down in some city or town, though just which one he could not guess, as the country below was wholly unfamiliar. Finally a small settlement loomed ahead. It was of the familiar style of toy-building-block architecture affected by the ant-men, and, from its appearance, was very new. On its outskirts further building operations were actively in progress. Apparently a few survivors of the accursed race of Formians were consolidating their position and attempting to build up a new empire in some out-of-the-way portion of the continent. As the earth-man was turning these thoughts over in his mind the plane softly settled down upon one of the flat roofs, and its occupants disembarked. Three of the ants advanced menacingly toward Myles, but Doggo held them off. Then all of the party descended down one of the ramps to the lower levels of the building. Narrow slitlike window openings gave onto courtyards, where fountains played and masses of blue and yellow flowers bloomed, amid gray-branched lichens with red and purple twig-knobs. It was in just such a garden, through just such a window, that he had first looked upon the lovely blue-eyed, golden-haired Lilla, Crown Princess of Cupia. The earth-man sighed. Where was his beloved wife now? That she needed his help was certain. He must therefore get busy. So once again he made motions of writing on the palm of his left hand with the thumb and forefinger of his right; and this time the sign language produced results, for Doggo halted the procession and led Cabot into a room. 13 It was a plain bare room, devoid of any furniture except a small table, for ant-men have no use for chairs and couches. The sky outside was already beginning to pinken with the unseen sun. With a sweep of his paw, Doggo indicated that this was to be Cabot’s quarters. Then, with another wave, he pointed to the table, where lay a pad of paper and stylus, not a pencil-like stylus as employed by the Cupians, but rather one equipped with straps for attaching it to the claw of a Formian. Even so, it was better than nothing. The earth-man seized it eagerly, but before he could begin writing an ant entered bearing a Cupian toga, short-sleeved and bordered with Grecian wave designs in blue. Myles put on this garment, and then quickly filled a sheet with questions: “How is my princess and my son, the baby king? Whence come all you Formians, whose race I thought had been exterminated? What part of Cupia is this? What is this city? Where is Prince Yuri? And what do you intend to do with me this time?” Then he passed the paper and stylus over to his old friend Doggo. They were alone together at last. The ant-man’s reply consumed sheet after sheet of paper; but, owning to the rapidity of Porovian shorthand, did not take so very much more time than speaking would have required. As he completed each sheet he passed it over to Myles, who read as follows: “As to your princess and your son, I know not, for this is not Cupia. Do you remember how, when your victorious army and air navy swept to the southern extremity of what had been Formia, a few of our survivors rose in planes from the ruins of our last stronghold and braved the dangers of the steam clouds which overhang the boiling seas? Our leader was Prince Yuri, erstwhile contender for the throne of Cupia, splendid even in defeat. “It was his brain that conceived our daring plan of escape. If there were other lands beyond the boiling seas, the lands which tradition taught were the origin of the Cupian race, then there we might prosper and raise up a new empire. At the worst we should merely meet death in another form, rather than at your hands. So we essayed. 14 “Your planes followed us, but turned back as we neared the area of terrific heat. Soon the vapor closed over us, blotting our enemies and our native land from view.” For page after page Doggo, the ant-man, related the harrowing details of that perilous flight across the boiling seas, ending with the words: “Here we are, and here are you, in Yuriana, capitol of New Formia. But how is it that you, Myles Cabot, have arrived here on this continent in exactly the same manner and condition in which I discovered you in old Formia eight years ago?” When Myles reached the end of reading this narrative, he in turn took the pad and stylus and related how he had gone to the planet Minos (which we call the Earth) to learn the latest discoveries and inventions there, and how his calculations for his return to Poros had been upset by some static conditions just as he had been about to transmit himself back. Oh, if only he had landed by chance upon the same beach as on his first journey through the skies! Wisely he refrained from mentioning the “S O S” message from Lilla. But his recollection of her predicament spurred him to be anxious about her rescue. His immediate problem was to learn what the ant-men planned for him; so the concluding words which he wrote upon the pad were: “And, now that you have me in your power, what shall you do with me?” “Old friend,” Doggo wrote in reply, “that depends entirely upon Yuri, our king, whose toga you now have on.” III YURI OR FORMIS? The earth-man grimaced, but then smiled. Perhaps, his succeeding to the toga of King Yuri might prove to be an omen. 15 “So Yuri is king of the ants?” he asked. “Yes,” his captor replied, “for Queen Formis did not survive the trip across the boiling seas.” “Then what of your empire?” Myles inquired. “No queen. No eggs. How can your race continue? For you Formians are like the ants on my own planet Minos.” Doggo’s reply astounded him. “Do you remember back at Wautoosa, I told you that some of us lesser Formians had occasionally laid eggs? So now behold before you Doggo, Admiral of the Formian Air Navy, and mother of a new Queen Formis.” This was truly a surprise! All along Cabot had always regarded the Formians as mannish. And rightly so, for they performed in their own country the duties assigned to men among the Cupians. Furthermore, all Formians, save only the reigning Formis herself, were called by the Porovian pronoun, which corresponds to “he” in English. When Myles had somewhat recovered from his astonishment, he warmly congratulated his friend by patting him on the side of the head, as is the Porovian custom. “Doggo,” he wrote, “this ought to constitute you a person of some importance among the Formians.” “It ought to,” the ant-man replied, “but as a matter of fact, it merely intensifies Yuri’s mistrust and hatred of me. Now that I am mother of the queen, he fears that I may turn against him and establish Formis in his place as the head of an empire of the Formians, by the Formians, and for the Formians exclusively.” “Why don’t you?” Myles wrote. It seemed to him to be a bully good idea, and incidentally a solution of his own difficulties. But Doggo wrote in horror, “It would be treason!” Then tore up all the correspondence. It is difficult to inculcate the thought of independence in the mind of one reared in an autocracy. The earth-man, however, persisted. “How many of the council can you count on, if the interests of Yuri should clash with those of Formis?” 16 “Only one—myself.” And again Doggo tore up the correspondence. Myles tactfully changed the subject. “Where is the arch-fiend now?” he asked. “We know not,” the Formian wrote in reply. “Six days ago he left us in his airship and flew westward. When he failed to return, we sent out scout planes to search for him, and we have been hunting ever since. When we sighted you on the beach this morning we thought that you might be our lost leader, and that is why we landed and approached you.” At about this point the conversation was interrupted by a worker ant who brought food: roast alta and green aphid milk. With what relish did the earth-man plunge into the feast, his first taste of Porovian delicacies in many months. During the meal conversation lagged, owing to the difficulty of writing and eating at the same time. But now Myles Cabot seized his pad and stylus and wrote: “Have you ever known me to fail in any undertaking on the planet Poros?” “No,” the ant-man wrote in reply. “Have you ever known me to be untrue to a principle, a cause, or a friend?” “No,” Doggo replied. “Then,” Myles wrote, “let us make your daughter queen in fact as well as in name.” “It is treason,” Doggo wrote in reply, but this time he did not tear up the correspondence. “Treason?” Myles asked. If he had spoken the word, he would have spoken it with scorn and derision. “Treason? Is it treason to support your own queen? What has become of the national pride of the once great Formians? Look! I pledge myself to the cause of Formis, rightful Queen of Formia. Formis, daughter of Doggo! What say you?” This time, as he tore up the correspondence, Doggo signified an affirmative. And thus there resulted further correspondence. 17 “Doggo,” Myles wrote, “can you get to the antenna of the queen?” The ant-man indicated that he could. “If she has inherited any of your character,” Myles continued, “she will assert herself, if given half a chance.” So the Pitmanesque conversation continued. Long since had the pink light of Porovian evening faded from the western sky. The ceiling vapor-lamps were lit. The night showed velvet-black through the slit-like windows. And still the two old friends wrote on, Myles Standish Cabot, the Bostonian, and Doggo, No. 334-2-18, the only really humanlike ant-man whom Myles had ever known among the once dominant race of Poros. Finally, as the dials indicated midnight, the two conspirators ceased their labors. All was arranged for the coup d’ etat . They tore into shreds every scrap of used paper, leaving extant merely the ant-man’s concluding words: “Meanwhile you are my prisoner.” Doggo then rang a soundless bell, which was answered by a worker ant, whom he inaudibly directed to bring sufficient draperies to form a bed for the earth-man. These brought, the two friends patted each other a fond good night, and the tired earth-man lay down for the first sleep which he had had in over forty earth hours. It hardly seemed possible! Night before last he had slept peacefully on a conventional feather-bed in a little New England farmhouse. Then had come the S O S message from the skies; and here he was now, millions of miles away through space retiring on matted silver felting on the concrete floor of a Porovian ant-house. Such are the mutations of fortune! With these thoughts the returned wanderer lapsed into a deep and dreamless sleep. When he awakened in the morning there was a guard posted at the door. 18 Doggo did not show up until nearly noon, when he rattled in, bristling with excitement. Seizing the pad he wrote: “A stormy session of the Council of Twelve! We are all agreed that you must be indicted for high crimes and misdemeanors. But the great question is as to just what we can charge you with.” “Sorry I can’t assist you,” the earth-man wrote. “How would it be if I were to slap your daughter’s face, or something? Or why not try me for general cussedness?” “That is just what we finally decided to do,” the ant-man wrote in reply. “We shall try you on general principles, and let the proper accusation develop from the evidence. “At some stage of the proceedings it will inevitably occur to some member of the council to suggest that you be charged with treason to Yuri, whereupon two members of the council, whom I have won over to the cause of my daughter, will raise the objection that Yuri is not our king. This will be the signal for the proclaiming of Queen Formis. If you will waive counsel the trial can take place to-morrow.” “I will waive anything,” Myles replied, “counsel, immunity, extradition, anything in order to speed up my return to Cupia, where Lilla awaits in some dire extremity.” “All right,” Doggo wrote, and the conference was at an end. The morrow would decide the ascendancy of Myles Cabot or the Prince Yuri over the new continent. IV THE COUP D’ETAT The next morning Myles Cabot was led under guard to the council chamber of the dread thirteen: Formis and her twelve advisers. The accused was placed in a wicker cage, from which he surveyed his surroundings as the proceedings opened. 19 On a raised platform stood the ant queen, surmounted by a scarlet canopy, which set off the perfect proportions of her jet-black body. On each side of her stood six refined and intelligent ant-men, her councillors. One of the twelve was Doggo. Messenger ants hurried hither and thither. First the accusation was read, Myles being furnished with a written copy. The witnesses were then called. They were veterans who had served in the wars in which Cabot had twice freed Cupia from the domination of its Formian oppressors. They spoke with bitterness of the downfall of their beloved Formia. Their testimony was brief. Then the accused was asked if he wished to say anything in his own behalf. Myles rose, then shrugged his shoulders, sat down again, and wrote: “I fully realize the futility of making an argument through the antennae of another.” Whereupon the queen and the council went into executive session. Their remarks were not intended for the eyes of the prisoner, but he soon observed that some kind of a dispute was on between Doggo, supported by two councillors named Emu and Fum on one side, and a councillor named Barth on the other. As this dispute reached its height, a messenger ant rushed in and held up one paw. Cabot’s interpreter, not deeming this a part of the executive session, obligingly translated the following into writing: The messenger: “Yuri lives and reigns over Cupia. It is his command that Cabot die.” Barth: “It is the radio. Know then, O Queen, and ye, members of the council, that when we fled across the boiling seas under the gallant leadership of Prince Yuri, the man with the heart of a Formian, he brought with him one of those powerful radio sets invented by the beast who is our prisoner here to-day. “Supporters of Yuri still remained among the Cupians, and he has been in constant communication with these ever since shortly after our arrival here. From them he learned of the return of Myles Cabot to the planet Minos. 20 “Then Yuri disappeared. Those of us who were closest to him suspected that he had gone back across the boiling seas to claim as his own the throne of Cupia. But we hesitated to announce this until we were sure, for we feared that some of our own people would regard his departure as desertion. Yet who can blame him for returning to his father-land and to the throne which is his by rights?” To which the messenger added: “And he offers to give us back our own old country, if we too will return across the boiling seas again.” “It is a lie!” Doggo shouted. “Yuri, usurper of the thrones of two continents. Bah!” shouted Emu. “Yuri, our rightful leader,” shouted Barth. “Give us a queen of our own race,” shouted Fum. “Release the prisoner,” shouted the Queen. And that is all that Myles learned of the conversation, for his interpreter at this juncture stopped writing and obeyed the queen. The earth-man was free! With one bound he gained the throne, where fighting was already in progress between the two factions. Barth and Doggo were rolling over and over on the floor in a death grapple, while the ant-queen had backed to the rear of the stage, closely guarded by Emu and Fum. Seizing one of the pikes which supported the scarlet canopy, Myles wrenched it loose and drove it into the thorax of Barth. In another instant the earth-man and Doggo stood beside the queen. Ant-men now came pouring into the chamber through all the entrances, taking sides as they entered and sized up the situation. If it had still been in vogue among the Formians to be known by numbers rather than names, and to have these identifying numbers painted on the backs of their abdomens followed by the numbers of those whom they had defeated in the duels so common among them, then many a Formian would have “got the number” of many another, that day. | D. Via antenna |
what were the baselines? | ### None
[block]I.1em [block]i.1em Learning to Rank Scientific Documents from the CrowdLearning to Rank Scientific Documents from the Crowd -4 [1]1 ### Introduction
The number of biomedical research papers published has increased dramatically in recent years. As of October, 2016, PubMed houses over 26 million citations, with almost 1 million from the first 3 quarters of 2016 alone . It has become impossible for any one person to actually read all of the work being published. We require tools to help us determine which research articles would be most informative and related to a particular question or document. For example, a common task when reading articles is to find articles that are most related to another. Major research search engines offer such a “related articles” feature. However, we propose that instead of measuring relatedness by text-similarity measures, we build a model that is able to infer relatedness from the authors' judgments. BIBREF0 consider two kinds of queries important to bibliographic information retrieval: the first is a search query written by the user and the second is a request for documents most similar to a document already judged relevant by the user. Such a query-by-document (or query-by-example) system has been implemented in the de facto scientific search engine PubMed—called Related Citation Search. BIBREF1 show that 19% of all PubMed searches performed by users have at least one click on a related article. Google Scholar provides a similar Related Articles system. Outside of bibliographic retrieval, query-by-document systems are commonly used for patent retrieval, Internet search, and plagiarism detection, amongst others. Most work in the area of query-by-document uses text-based similarity measures ( BIBREF2 , BIBREF3 , BIBREF4 ). However, scientific research is hypothesis driven and therefore we question whether text-based similarity alone is the best model for bibliographic retrieval. In this study we asked authors to rank documents by “closeness” to their work. The definition of “closeness” was left for the authors to interpret, as the goal is to model which documents the authors subjectively feel are closest to their own. Throughout the paper we will use “closeness” and “relatedness” interchangeably. We found that researchers' ranking by closeness differs significantly from the ranking provided by a traditional IR system. Our contributions are three fold: The principal ranking algorithms of query-by-document in bibliographic information retrieval rely mainly on text similarity measures ( BIBREF1 , BIBREF0 ). For example, the foundational work of BIBREF0 introduced the concept of a “document neighborhood” in which they pre-compute a text-similarity based distance between each pair of documents. When a user issues a query, first an initial set of related documents is retrieved. Then, the neighbors of each of those documents is retrieved, i.e., documents with the highest text similarity to those in the initial set. In a later work, BIBREF1 develop the PMRA algorithm for PubMed related article search. PMRA is an unsupervised probabilistic topic model that is trained to model “relatedness” between documents. BIBREF5 introduce the competing algorithm Find-Similar for this task, treating the full text of documents as a query and selecting related documents from the results. Outside bibliographic IR, prior work in query-by-document includes patent retrieval ( BIBREF6 , BIBREF3 ), finding related documents given a manuscript ( BIBREF1 , BIBREF7 ), and web page search ( BIBREF8 , BIBREF9 ). Much of the work focuses on generating shorter queries from the lengthy document. For example, noun-phrase extraction has been used for extracting short, descriptive phrases from the original lengthy text ( BIBREF10 ). Topic models have been used to distill a document into a set of topics used to form query ( BIBREF11 ). BIBREF6 generated queries using the top TF*IDF weighted terms in each document. BIBREF4 suggested extracting phrasal concepts from a document, which are then used to generate queries. BIBREF2 combined query extraction and pseudo-relevance feedback for patent retrieval. BIBREF9 employ supervised machine learning model (i.e., Conditional Random Fields) ( BIBREF12 ) for query generation. BIBREF13 explored ontology to identify chemical concepts for queries. There are also many biomedical-document specific search engines available. Many information retrieval systems focus on question answering systems such as those developed for the TREC Genomics Track ( BIBREF14 ) or BioASQ Question-Answer ( BIBREF15 ) competitions. Systems designed for question-answering use a combination of natural language processing techniques to identify biomedical entities, and then information retrieval systems to extract relevant answers to questions. Systems like those detailed in BIBREF16 can provide answers to yes/no biomedical questions with high precision. However what we propose differs from these systems in a fundamental way: given a specific document, suggest the most important documents that are related to it. The body of work most related to ours is that of citation recommendation. The goal of citation recommendation is to suggest a small number of publications that can be used as high quality references for a particular article ( BIBREF17 , BIBREF1 ). Topic models have been used to rank articles based on the similarity of latent topic distribution ( BIBREF11 , BIBREF18 , BIBREF1 ). These models attempt to decompose a document into a few important keywords. Specifically, these models attempt to find a latent vector representation of a document that has a much smaller dimensionality than the document itself and compare the reduced dimension vectors. Citation networks have also been explored for ranking articles by importance, i.e., authority ( BIBREF19 , BIBREF20 ). BIBREF17 introduced heterogeneous network models, called meta-path based models, to incorporate venues (the conference where a paper is published) and content (the term which links two articles, for citation recommendation). Another highly relevant work is BIBREF8 who decomposed a document to represent it with a compact vector, which is then used to measure the similarity with other documents. Note that we exclude the work of context-aware recommendation, which analyze each citation's local context, which is typically short and does not represent a full document. One of the key contributions of our study is an innovative approach for automatically generating a query-by-document gold standard. Crowd-sourcing has generated large databases, including Wikipedia and Freebase. Recently, BIBREF21 concluded that unpaid participants performed better than paid participants for question answering. They attribute this to unpaid participants being more intrinsically motivated than the paid test takers: they performed the task for fun and already had knowledge about the subject being tested. In contrast, another study, BIBREF22 , compared unpaid workers found through Google Adwords (GA) to paid workers found through Amazon Mechanical Turk (AMT). They found that the paid participants from AMT outperform the unpaid ones. This is attributed to the paid workers being more willing to look up information they didn't know. In the bibliographic domain, authors of scientific publications have contributed annotations ( BIBREF23 ). They found that authors are more willing to annotate their own publications ( BIBREF23 ) than to annotate other publications ( BIBREF24 ) even though they are paid. In this work, our annotated dataset was created by the unpaid authors of the articles. ### Benchmark Datasets
In order to develop and evaluate ranking algorithms we need a benchmark dataset. However, to the best of our knowledge, we know of no openly available benchmark dataset for bibliographic query-by-document systems. We therefore created such a benchmark dataset. The creation of any benchmark dataset is a daunting labor-intensive task, and in particular, challenging in the scientific domain because one must master the technical jargon of a scientific article, and such experts are not easy to find when using traditional crowd-sourcing technologies (e.g., AMT). For our task, the ideal annotator for each of our articles are the authors themselves. The authors of a publication typically have a clear knowledge of the references they cite and their scientific importance to their publication, and therefore may be excellent judges for ranking the reference articles. Given the full text of a scientific publication, we want to rank its citations according to the author's judgments. We collected recent publications from the open-access PLoS journals and asked the authors to rank by closeness five citations we selected from their paper. PLoS articles were selected because its journals cover a wide array of topics and the full text articles are available in XML format. We selected the most recent publications as previous work in crowd-sourcing annotation shows that authors' willingness to participate in an unpaid annotation task declines with the age of publication ( BIBREF23 ). We then extracted the abstract, citations, full text, authors, and corresponding author email address from each document. The titles and abstracts of the citations were retrieved from PubMed, and the cosine similarity between the PLoS abstract and the citation's abstract was calculated. We selected the top five most similar abstracts using TF*IDF weighted cosine similarity, shuffled their order, and emailed them to the corresponding author for annotation. We believe that ranking five articles (rather than the entire collection of the references) is a more manageable task for an author compared to asking them to rank all references. Because the documents to be annotated were selected based on text similarity, they also represent a challenging baseline for models based on text-similarity features. In total 416 authors were contacted, and 92 responded (22% response rate). Two responses were removed from the dataset for incomplete annotation. We asked authors to rank documents by how “close to your work” they were. The definition of closeness was left to the discretion of the author. The dataset is composed of 90 annotated documents with 5 citations each ranked 1 to 5, where 1 is least relevant and 5 is most relevant for a total of 450 annotated citations. ### Learning to Rank
Learning-to-rank is a technique for reordering the results returned from a search engine query. Generally, the initial query to a search engine is concerned more with recall than precision: the goal is to obtain a subset of potentially related documents from the corpus. Then, given this set of potentially related documents, learning-to-rank algorithms reorder the documents such that the most relevant documents appear at the top of the list. This process is illustrated in Figure FIGREF6 . There are three basic types of learning-to-rank algorithms: point-wise, pair-wise, and list-wise. Point-wise algorithms assign a score to each retrieved document and rank them by their scores. Pair-wise algorithms turn learning-to-rank into a binary classification problem, obtaining a ranking by comparing each individual pair of documents. List-wise algorithms try to optimize an evaluation parameter over all queries in the dataset. Support Vector Machine (SVM) ( BIBREF25 ) is a commonly used supervised classification algorithm that has shown good performance over a range of tasks. SVM can be thought of as a binary linear classifier where the goal is to maximize the size of the gap between the class-separating line and the points on either side of the line. This helps avoid over-fitting on the training data. SVMRank is a modification to SVM that assigns scores to each data point and allows the results to be ranked ( BIBREF26 ). We use SVMRank in the experiments below. SVMRank has previously been used in the task of document retrieval in ( BIBREF27 ) for a more traditional short query task and has been shown to be a top-performing system for ranking. SVMRank is a point-wise learning-to-rank algorithm that returns scores for each document. We rank the documents by these scores. It is possible that sometimes two documents will have the same score, resulting in a tie. In this case, we give both documents the same rank, and then leave a gap in the ranking. For example, if documents 2 and 3 are tied, their ranked list will be [5, 3, 3, 2, 1]. Models are trained by randomly splitting the dataset into 70% training data and 30% test data. We apply a random sub-sampling approach where the dataset is randomly split, trained, and tested 100 times due to the relatively small size of the data. A model is learned for each split and a ranking is produced for each annotated document. We test three different supervised models. The first supervised model uses only text similarity features, the second model uses all of the features, and the third model runs forward feature selection to select the best performing combination of features. We also test using two different models trained on two different datasets: one trained using the gold standard annotations, and another trained using the judgments based on text similarity that were used to select the citations to give to the authors. We tested several different learning to rank algorithms for this work. We found in preliminary testing that SVMRank had the best performance, so it will be used in the following experiments. ### Features
Each citation is turned into a feature vector representing the relationship between the published article and the citation. Four types of features are used: text similarity, citation count and location, age of the citation, and the number of times the citation has appeared in the literature (citation impact). Text similarity features measure the similarity of the words used in different parts of the document. In this work, we calculate the similarity between a document INLINEFORM0 and a document it cites INLINEFORM1 by transforming the their text into term vectors. For example, to calculate the similarity of the abstracts between INLINEFORM2 and INLINEFORM3 we transform the abstracts into two term vectors, INLINEFORM4 and INLINEFORM5 . The length of each of the term vectors is INLINEFORM6 . We then weight each word by its Term-frequency * Inverse-document frequency (TF*IDF) weight. TF*IDF is a technique to give higher weight to words that appear frequently in a document but infrequently in the corpus. Term frequency is simply the number of times that a word INLINEFORM7 appears in a document. Inverse-document frequency is the logarithmically-scaled fraction of documents in the corpus in which the word INLINEFORM8 appears. Or, more specifically: INLINEFORM9 where INLINEFORM0 is the total number of documents in the corpus, and the denominator is the number of documents in which a term INLINEFORM1 appears in the corpus INLINEFORM2 . Then, TF*IDF is defined as: INLINEFORM3 where INLINEFORM0 is a term, INLINEFORM1 is the document, and INLINEFORM2 is the corpus. For example, the word “the” may appear often in a document, but because it also appears in almost every document in the corpus it is not useful for calculating similarity, thus it receives a very low weight. However, a word such as “neurogenesis” may appear often in a document, but does not appear frequently in the corpus, and so it receives a high weight. The similarity between term vectors is then calculated using cosine similarity: INLINEFORM3 where INLINEFORM0 and INLINEFORM1 are two term vectors. The cosine similarity is a measure of the angle between the two vectors. The smaller the angle between the two vectors, i.e., the more similar they are, then the closer the value is to 1. Conversely, the more dissimilar the vectors, the closer the cosine similarity is to 0. We calculate the text similarity between several different sections of the document INLINEFORM0 and the document it cites INLINEFORM1 . From the citing article INLINEFORM2 , we use the title, full text, abstract, the combined discussion/conclusion sections, and the 10 words on either side of the place in the document where the actual citation occurs. From the document it cites INLINEFORM3 we only use the title and the abstract due to limited availability of the full text. In this work we combine the discussion and conclusion sections of each document because some documents have only a conclusion section, others have only a discussion, and some have both. The similarity between each of these sections from the two documents is calculated and used as features in the model. The age of the citation may be relevant to its importance. As a citation ages, we hypothesize that it is more likely to become a “foundational” citation rather than one that directly influenced the development of the article. Therefore more recent citations may be more likely relevant to the article. Similarly, “citation impact”, that is, the number of times a citation has appeared in the literature (as measured by Google Scholar) may be an indicator of whether or not an article is foundational rather than directly related. We hypothesize that the fewer times an article is cited in the literature, the more impact it had on the article at hand. We also keep track of the number of times a citation is mentioned in both the full text and discussion/conclusion sections. We hypothesize that if a citation is mentioned multiple times, it is more important than citations that are mentioned only once. Further, citations that appear in the discussion/conclusion sections are more likely to be crucial to understanding the results. We normalize the counts of the citations by the total number of citations in that section. In total we select 15 features, shown in Table TABREF15 . The features are normalized within each document so that each of citation features is on a scale from 0 to 1, and are evenly distributed within that range. This is done because some of the features (such as years since citation) are unbounded. ### Baseline Systems
We compare our system to a variety of baselines. (1) Rank by the number of times a citation is mentioned in the document. (2) Rank by the number of times the citation is cited in the literature (citation impact). (3) Rank using Google Scholar Related Articles. (4) Rank by the TF*IDF weighted cosine similarity. (5) Rank using a learning-to-rank model trained on text similarity rankings. The first two baseline systems are models where the values are ordered from highest to lowest to generate the ranking. The idea behind them is that the number of times a citation is mentioned in an article, or the citation impact may already be good indicators of their closeness. The text similarity model is trained using the same features and methods used by the annotation model, but trained using text similarity rankings instead of the author's judgments. We also compare our rankings to those found on the popular scientific article search engine Google Scholar. Google Scholar is a “black box” IR system: they do not release details about which features they are using and how they judge relevance of documents. Google Scholar provides a “Related Articles” feature for each document in its index that shows the top 100 related documents for each article. To compare our rankings, we search through these related documents and record the ranking at which each of the citations we selected appeared. We scale these rankings such that the lowest ranked article from Google Scholar has the highest relevance ranking in our set. If the cited document does not appear in the set, we set its relevance-ranking equal to one below the lowest relevance ranking found. Four comparisons are performed with the Google Scholar data. (1) We first train a model using our gold standard and see if we can predict Google Scholar's ranking. (2) We compare to a baseline of using Google Scholar's rankings to train and compare with their own rankings using our feature set. (3) Then we train a model using Google Scholar's rankings and try to predict our gold standard. (4) We compare it to the model trained on our gold standard to predict our gold standard. ### Evaluation Measures
Normalized Discounted Cumulative Gain (NDCG) is a common measure for comparing a list of estimated document relevance judgments with a list of known judgments ( BIBREF28 ). To calculate NDCG we first calculate a ranking's Discounted Cumulative Gain (DCG) as: DISPLAYFORM0 where rel INLINEFORM0 is the relevance judgment at position INLINEFORM1 . Intuitively, DCG penalizes retrieval of documents that are not relevant (rel INLINEFORM2 ). However, DCG is an unbounded value. In order to compare the DCG between two models, we must normalize it. To do this, we use the ideal DCG (IDCG), i.e., the maximum possible DCG given the relevance judgments. The maximum possible DCG occurs when the relevance judgments are in the correct order. DISPLAYFORM0 The NDCG value is in the range of 0 to 1, where 0 means that no relevant documents were retrieved, and 1 means that the relevant documents were retrieved and in the correct order of their relevance judgments. Kendall's INLINEFORM0 is a measure of the correlation between two ranked lists. It compares the number of concordant pairs with the number of discordant pairs between each list. A concordant pair is defined over two observations INLINEFORM1 and INLINEFORM2 . If INLINEFORM3 and INLINEFORM4 , then the pair at indices INLINEFORM5 is concordant, that is, the ranking at INLINEFORM6 in both ranking sets INLINEFORM7 and INLINEFORM8 agree with each other. Similarly, a pair INLINEFORM9 is discordant if INLINEFORM10 and INLINEFORM11 or INLINEFORM12 and INLINEFORM13 . Kendall's INLINEFORM14 is then defined as: DISPLAYFORM0 where C is the number of concordant pairs, D is the number of discordant pairs, and the denominator represents the total number of possible pairs. Thus, Kendall's INLINEFORM0 falls in the range of INLINEFORM1 , where -1 means that the ranked lists are perfectly negatively correlated, 0 means that they are not significantly correlated, and 1 means that the ranked lists are perfectly correlated. One downside of this measure is that it does not take into account where in the ranked list an error occurs. Information retrieval, in general, cares more about errors near the top of the list rather than errors near the bottom of the list. Average-Precision INLINEFORM0 ( BIBREF29 ) (or INLINEFORM1 ) extends on Kendall's INLINEFORM2 by incorporating the position of errors. If an error occurs near the top of the list, then that is penalized heavier than an error occurring at the bottom of the list. To achieve this, INLINEFORM3 incorporates ideas from the popular Average Precision measure, were we calculate the precision at each index of the list and then average them together. INLINEFORM4 is defined as: DISPLAYFORM0 Intuitively, if an error occurs at the top of the list, then that error is propagated into each iteration of the summation, meaning that it's penalty is added multiple times. INLINEFORM0 's range is between -1 and 1, where -1 means the lists are perfectly negatively correlated, 0 means that they are not significantly correlated, and 1 means that they are perfectly correlated. ### Forward Feature Selection
Forward feature selection was performed by iteratively testing each feature one at a time. The highest performing feature is kept in the model, and another sweep is done over the remaining features. This continues until all features have been selected. This approach allows us to explore the effect of combinations of features and the effect of having too many or too few features. It also allows us to evaluate which features and combinations of features are the most powerful. ### Results
We first compare our gold standard to the baselines. A random baseline is provided for reference. Because all of the documents that we rank are relevant, NDCG will be fairly high simply by chance. We find that the number of times a document is mentioned in the annotated document is significantly better than the random baseline or the citation impact. The more times a document is mentioned in a paper, the more likely the author was to annotate it as important. Interestingly, we see a negative correlation with the citation impact. The more times a document is mentioned in the literature, the less likely it is to be important. These results are shown in Table TABREF14 . Next we rank the raw values of the features and compare them to our gold standard to obtain a baseline (Table TABREF15 ). The best performing text similarity feature is the similarity between the abstract of the annotated document and the abstract of the cited document. However, the number of times that a cited document is mentioned in the text of the annotated document are also high-scoring features, especially in the INLINEFORM0 correlation coefficient. These results indicate that text similarity alone may not be a good measure for judging the rank of a document. Next we test three different feature sets for our supervised learning-to-rank models. The model using only the text similarity features performs poorly: NDCG stays at baseline and the correlation measures are low. Models that incorporate information about the age, number of times a cited document was referenced, and the citation impact of that document in addition to the text similarity features significantly outperformed models that used only text similarity features INLINEFORM0 . Because INLINEFORM1 takes into account the position in the ranking of the errors, this indicates that the All Features model was able to better correctly place highly ranked documents above lower ranked ones. Similarly, because Kendall's INLINEFORM2 is an overall measure of correlation that does not take into account the position of errors, the higher value here means that more rankings were correctly placed. Interestingly, feature selection (which is optimized for NDCG) does not outperform the model using all of the features in terms of our correlation measures. The features chosen during forward feature selection are (1) the citation impact, (2) number of mentions in the full text, (3) text similarity between the annotated document's title and the referenced document's abstract, (4) the text similarity between the annotated document's discussion/conclusion section and the referenced document's title. These results are shown in Table TABREF16 . The models trained on the text similarity judgments perform worse than the models trained on the annotated data. However, in terms of both NDCG and the correlation measures, they perform significantly better than the random baseline. Next we compare our model to Google Scholar's rankings. Using the ranking collected from Google Scholar, we build a training set to try to predict our authors' rankings. We find that Google Scholar performs similarly to the text-only features model. This indicates that the rankings we obtained from the authors are substantially different than the rankings that Google Scholar provides. Results appear in Table TABREF17 . ### Discussion
We found that authors rank the references they cite substantially differently from rankings based on text-similarity. Our results show that decomposing a document into a set of features that is able to capture that difference is key. While text similarity is indeed important (as evidenced by the Similarity(a,a) feature in Table TABREF15 ), we also found that the number of times a document is referenced in the text and the number of times a document is referenced in the literature are also both important features (via feature selection). The more often a citation is mentioned in the text, the more likely it is to be important. This feature is often overlooked in article citation recommendation. We also found that recency is important: the age of the citation is negatively correlated with the rank. Newer citations are more likely to be directly important than older, more foundational citations. Additionally, the number of times a document is cited in the literature is negatively correlated with rank. This is likely due to highly cited documents being more foundational works; they may be older papers that are important to the field but not directly influential to the new work. The model trained using the author's judgments does significantly better than the model trained using the text-similarity-based judgments. An error analysis was performed to find out why some of the rankings disagreed with the author's annotations. We found that in some cases our features were unable to capture the relationship: for example a biomedical document applying a model developed in another field to the dataset may use very different language to describe the model than the citation. Previous work adopting topic models to query document search may prove useful for such cases. A small subset of features ended up performing as well as the full list of features. The number of times a citation was mentioned and the citation impact score in the literature ended up being two of the most important features. Indeed, without the citation-based features, the model performs as though it were trained with the text-similarity rankings. Feature engineering is a part of any learning-to-rank system, especially in domain-specific contexts. Citations are an integral feature of our dataset. For learning-to-rank to be applied to other datasets feature engineering must also occur to exploit the unique properties of those datasets. However, we show that combining the domain-specific features with more traditional text-based features does improve the model's scores over simply using the domain-specific features themselves. Interestingly, citation impact and age of the citation are both negatively correlated with rank. We hypothesize that this is because both measures can be indicators of recency: a new publication is more likely to be directly influenced by more recent work. Many other related search tools, however, treat the citation impact as a positive feature of relatedness: documents with a higher citation impact appear higher on the list of related articles than those with lower citation impacts. This may be the opposite of what the user actually desires. We also found that rankings from our text-similarity based IR system or Google Scholar's IR system were unable to rank documents by the authors' annotations as well as our system. In one sense, this is reasonable: the rankings coming from these systems were from a different system than the author annotations. However, in domain-specific IR, domain experts are the best judges. We built a system that exploits these expert judgments. The text similarity and Google Scholar models were able to do this to some extent, performing above the random baseline, but not on the level of our model. Additionally, we observe that NDCG may not be the most appropriate measure for comparing short ranked lists where all of the documents are relevant to some degree. NDCG gives a lot of credit to relevant documents that occur in the highest ranks. However, all of the documents here are relevant, just to varying degrees. Thus, NDCG does not seem to be the most appropriate measure, as is evident in our scores. The correlation coefficients from Kendall's INLINEFORM0 and INLINEFORM1 seem to be far more appropriate for this case, as they are not concerned with relevance, only ranking. One limitation of our work is that we selected a small set of references based on their similarities to the article that cites them. Ideally, we would have had authors rank all of their citations for us, but this would have been a daunting task for authors to perform. We chose to use the Google Scholar dataset in order to attempt to mitigate this: we obtain a ranking for the set of references from a system that is also ranking many other documents. The five citations selected by TF*IDF weighted cosine similarity represent a “hard” gold standard: we are attempting to rank documents that are known to all be relevant by their nature, and have high similarity with the text. Additionally, there are plethora of other, more expensive features we could explore to improve the model. Citation network features, phrasal concepts, and topic models could all be used to help improve our results, at the cost of computational complexity. We have developed a model for fast related-document ranking based on crowd-sourced data. The model, data, and data collection software are all publicly available and can easily be used in future applications as an automatic search to help users find the most important citations given a particular document. The experimental setup is portable to other datasets with some feature engineering. We were able to identify that several domain-specific features were crucial to our model, and that we were able to improve on the results of simply using those features alone by adding more traditional features. Query-by-document is a complicated and challenging task. We provide an approach with an easily obtained dataset and a computationally inexpensive model. By working with biomedical researchers we were able to build a system that ranks documents in a quantitatively different way than previous systems, and to provide a tool that helps researchers find related documents. ### Acknowledgments
We would like to thank all of the authors who took the time to answer our citation ranking survey. This work is supported by National Institutes of Health with the grant number 1R01GM095476. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript. Figure 1: The basic pipeline of a learning-to-rank system. An initial set of results for a query is retrieved from a search engine, and then that subset is reranked. During the reranking phase new features may be extracted. Table 1: Results for the citation baselines. The number of times a citation is mentioned in the document is a better indicator of rank than the citation impact. Table 2: Results for ranking by each individual feature value. Similarity features are text similarity features. The first parameter is the section of text in the annotated document, the second parameter is the section of text in the referenced document. Here, “a” means abstract, “t” means title, “f” means full text, “c” means the 10 word window around a citation, “d” means the discussion/conclusion sections, and “cd” means 10 word windows around citations in the discussion/conclusion section. Age is the age of the referenced document, MentionCount is the number of times the annotated document mentions the referenced document in text, and CitationImpact is the number of documents that have cited the referenced document in the literature. Table 3: Results for the SVMRank models for three different combinations of features. “Text Only Features” are only the text similarity features. “Feature Selection” is the set of features found after running a forward feature selection algorithm. A “*” indicates statistical significance between the two models. Table 4: Results for the model trained using the Google Scholar Related Articles ranking. We find that building a model using Google Scholar’s Related Articles ranking to predict our authors’ rankings performs poorly compared to the other models. | (1) Rank by the number of times a citation is mentioned in the document., (2) Rank by the number of times the citation is cited in the literature (citation impact)., (3) Rank using Google Scholar Related Articles., (4) Rank by the TF*IDF weighted cosine similarity., (5) Rank using a learning-to-rank model trained on text similarity rankings. |
How are aspects identified in aspect extraction? | ### Affiliation
School of Computer Science and Engineering, Nanyang Technological University, Singapore ### Synonyms
Sentiment Analysis, Subjectivity Detection, Deep Learning Aspect Extraction, Polarity Distribution, Convolutional Neural Network. ### Glossary
Aspect : Feature related to an opinion target Convolution : features made of consecutive words BOW : Bag of Words NLP : Natural Language Processing CNN : Convolutional Neural Network LDA : Latent Dirichlet Allocation ### Definition
Subjectivity detection is the task of identifying objective and subjective sentences. Objective sentences are those which do not exhibit any sentiment. So, it is desired for a sentiment analysis engine to find and separate the objective sentences for further analysis e.g., polarity detection. In subjective sentences, opinions can often be expressed on one or multiple topics. Aspect extraction is a subtask of sentiment analysis that consists in identifying opinion targets in opinionated text, i.e., in detecting the specific aspects of a product or service the opinion holder is either praising or complaining about. ### Key Points
We consider deep convolutional neural networks where each layer is learned independent of the others resulting in low complexity. We model temporal dynamics in product reviews by pre-training the deep CNN using dynamic Gaussian Bayesian networks. We combine linguistic aspect mining with CNN features for effective sentiment detection. ### Historical Background
Traditional methods prior to 2001 used hand-crafted templates to identify subjectivity and did not generalize well for resource-deficient languages such as Spanish. Later works published between 2002 and 2009 proposed the use of deep neural networks to automatically learn a dictionary of features (in the form of convolution kernels) that is portable to new languages. Recently, recurrent deep neural networks are being used to model alternating subjective and objective sentences within a single review. Such networks are difficult to train for a large vocabulary of words due to the problem of vanishing gradients. Hence, in this chapter we consider use of heuristics to learn dynamic Gaussian networks to select significant word dependencies between sentences in a single review. Further, in order to relation between opinion targets and the corresponding polarity in a review, aspect based opinion mining is used. Explicit aspects were models by several authors using statistical observations such mutual information between noun phrase and the product class. However this method was unable to detect implicit aspects due to high level of noise in the data. Hence, topic modeling was widely used to extract and group aspects, where the latent variable 'topic' is introduced between the observed variables 'document' and 'word'. In this chapter, we demonstrate the use of 'common sense reasoning' when computing word distributions that enable shifting from a syntactic word model to a semantic concept model. ### Introduction
While sentiment analysis research has become very popular in the past ten years, most companies and researchers still approach it simply as a polarity detection problem. In reality, sentiment analysis is a `suitcase problem' that requires tackling many natural language processing (NLP) subtasks, including microtext analysis, sarcasm detection, anaphora resolution, subjectivity detection and aspect extraction. In this chapter, we focus on the last two subtasks as they are key for ensuring a minimum level of accuracy in the detection of polarity from social media. The two basic issues associated with sentiment analysis on the Web, in fact, are that (1) a lot of factual or non-opinionated information needs to be filtered out and (2) opinions are most times about different aspects of the same product or service rather than on the whole item and reviewers tend to praise some and criticize others. Subjectivity detection, hence, ensures that factual information is filtered out and only opinionated information is passed on to the polarity classifier and aspect extraction enables the correct distribution of polarity among the different features of the opinion target (in stead of having one unique, averaged polarity assigned to it). In this chapter, we offer some insights about each task and apply an ensemble of deep learning and linguistics to tackle both. The opportunity to capture the opinion of the general public about social events, political movements, company strategies, marketing campaigns, and product preferences has raised increasing interest of both the scientific community (because of the exciting open challenges) and the business world (because of the remarkable benefits for marketing and financial market prediction). Today, sentiment analysis research has its applications in several different scenarios. There are a good number of companies, both large- and small-scale, that focus on the analysis of opinions and sentiments as part of their mission BIBREF0 . Opinion mining techniques can be used for the creation and automated upkeep of review and opinion aggregation websites, in which opinions are continuously gathered from the Web and not restricted to just product reviews, but also to broader topics such as political issues and brand perception. Sentiment analysis also has a great potential as a sub-component technology for other systems. It can enhance the capabilities of customer relationship management and recommendation systems; for example, allowing users to find out which features customers are particularly interested in or to exclude items that have received overtly negative feedback from recommendation lists. Similarly, it can be used in social communication for troll filtering and to enhance anti-spam systems. Business intelligence is also one of the main factors behind corporate interest in the field of sentiment analysis BIBREF1 . Sentiment analysis is a `suitcase' research problem that requires tackling many NLP sub-tasks, including semantic parsing BIBREF2 , named entity recognition BIBREF3 , sarcasm detection BIBREF4 , subjectivity detection and aspect extraction. In opinion mining, different levels of analysis granularity have been proposed, each one having its own advantages and drawbacks BIBREF5 , BIBREF6 . Aspect-based opinion mining BIBREF7 , BIBREF8 focuses on the relations between aspects and document polarity. An aspect, also known as an opinion target, is a concept in which the opinion is expressed in the given document. For example, in the sentence, “The screen of my phone is really nice and its resolution is superb” for a phone review contains positive polarity, i.e., the author likes the phone. However, more specifically, the positive opinion is about its screen and resolution; these concepts are thus called opinion targets, or aspects, of this opinion. The task of identifying the aspects in a given opinionated text is called aspect extraction. There are two types of aspects defined in aspect-based opinion mining: explicit aspects and implicit aspects. Explicit aspects are words in the opinionated document that explicitly denote the opinion target. For instance, in the above example, the opinion targets screen and resolution are explicitly mentioned in the text. In contrast, an implicit aspect is a concept that represents the opinion target of an opinionated document but which is not specified explicitly in the text. One can infer that the sentence, “This camera is sleek and very affordable” implicitly contains a positive opinion of the aspects appearance and price of the entity camera. These same aspects would be explicit in an equivalent sentence: “The appearance of this camera is sleek and its price is very affordable.” Most of the previous works in aspect term extraction have either used conditional random fields (CRFs) BIBREF9 , BIBREF10 or linguistic patterns BIBREF7 , BIBREF11 . Both of these approaches have their own limitations: CRF is a linear model, so it needs a large number of features to work well; linguistic patterns need to be crafted by hand, and they crucially depend on the grammatical accuracy of the sentences. In this chapter, we apply an ensemble of deep learning and linguistics to tackle both the problem of aspect extraction and subjectivity detection. The remainder of this chapter is organized as follows: Section SECREF3 and SECREF4 propose some introductory explanation and some literature for the tasks of subjectivity detection and aspect extraction, respectively; Section SECREF5 illustrates the basic concepts of deep learning adopted in this work; Section SECREF6 describes in detail the proposed algorithm; Section SECREF7 shows evaluation results; finally, Section SECREF9 concludes the chapter. ### Subjectivity detection
Subjectivity detection is an important subtask of sentiment analysis that can prevent a sentiment classifier from considering irrelevant or potentially misleading text in online social platforms such as Twitter and Facebook. Subjective extraction can reduce the amount of review data to only 60 INLINEFORM0 and still produce the same polarity results as full text classification BIBREF12 . This allows analysts in government, commercial and political domains who need to determine the response of people to different crisis events BIBREF12 , BIBREF13 , BIBREF14 . Similarly, online reviews need to be summarized in a manner that allows comparison of opinions, so that a user can clearly see the advantages and weaknesses of each product merely with a single glance, both in unimodal BIBREF15 and multimodal BIBREF16 , BIBREF17 contexts. Further, we can do in-depth opinion assessment, such as finding reasons or aspects BIBREF18 in opinion-bearing texts. For example, INLINEFORM1 , which makes the film INLINEFORM2 . Several works have explored sentiment composition through careful engineering of features or polarity shifting rules on syntactic structures. However, sentiment accuracies for classifying a sentence as positive/negative/neutral has not exceeded 60 INLINEFORM3 . Early attempts used general subjectivity clues to generate training data from un-annotated text BIBREF19 . Next, bag-of-words (BOW) classifiers were introduced that represent a document as a multi set of its words disregarding grammar and word order. These methods did not work well on short tweets. Co-occurrence matrices also were unable to capture difference in antonyms such as `good/bad' that have similar distributions. Subjectivity detection hence progressed from syntactic to semantic methods in BIBREF19 , where the authors used extraction pattern to represent subjective expressions. For example, the pattern `hijacking' of INLINEFORM0 , looks for the noun `hijacking' and the object of the preposition INLINEFORM1 . Extracted features are used to train machine-learning classifiers such as SVM BIBREF20 and ELM BIBREF21 . Subjectivity detection is also useful for constructing and maintaining sentiment lexicons, as objective words or concepts need to be omitted from them BIBREF22 . Since, subjective sentences tend to be longer than neutral sentences, recursive neural networks were proposed where the sentiment class at each node in the parse tree was captured using matrix multiplication of parent nodes BIBREF23 , BIBREF24 . However, the number of possible parent composition functions is exponential, hence in BIBREF25 recursive neural tensor network was introduced that use a single tensor composition function to define multiple bilinear dependencies between words. In BIBREF26 , the authors used logistic regression predictor that defines a hyperplane in the word vector space where a word vectors positive sentiment probability depends on where it lies with respect to this hyperplane. However, it was found that while incorporating words that are more subjective can generally yield better results, the performance gain by employing extra neutral words is less significant BIBREF27 . Another class of probabilistic models called Latent Dirichlet Allocation assumes each document is a mixture of latent topics. Lastly, sentence-level subjectivity detection was integrated into document-level sentiment detection using graphs where each node is a sentence. The contextual constraints between sentences in a graph led to significant improvement in polarity classification BIBREF28 . Similarly, in BIBREF29 the authors take advantage of the sequence encoding method for trees and treat them as sequence kernels for sentences. Templates are not suitable for semantic role labeling, because relevant context might be very far away. Hence, deep neural networks have become popular to process text. In word2vec, for example, a word's meaning is simply a signal that helps to classify larger entities such as documents. Every word is mapped to a unique vector, represented by a column in a weight matrix. The concatenation or sum of the vectors is then used as features for prediction of the next word in a sentence BIBREF30 . Related words appear next to each other in a INLINEFORM0 dimensional vector space. Vectorizing them allows us to measure their similarities and cluster them. For semantic role labeling, we need to know the relative position of verbs, hence the features can include prefix, suffix, distance from verbs in the sentence etc. However, each feature has a corresponding vector representation in INLINEFORM1 dimensional space learned from the training data. Recently, convolutional neural network (CNN) is being used for subjectivity detection. In particular, BIBREF31 used recurrent CNNs. These show high accuracy on certain datasets such as Twitter we are also concerned with a specific sentence within the context of the previous discussion, the order of the sentences preceding the one at hand results in a sequence of sentences also known as a time series of sentences BIBREF31 . However, their model suffers from overfitting, hence in this work we consider deep convolutional neural networks, where temporal information is modeled via dynamic Gaussian Bayesian networks. ### Aspect-Based Sentiment Analysis
Aspect extraction from opinions was first studied by BIBREF7 . They introduced the distinction between explicit and implicit aspects. However, the authors only dealt with explicit aspects and used a set of rules based on statistical observations. Hu and Liu's method was later improved by BIBREF32 and by BIBREF33 . BIBREF32 assumed the product class is known in advance. Their algorithm detects whether a noun or noun phrase is a product feature by computing the point-wise mutual information between the noun phrase and the product class. BIBREF34 presented a method that uses language model to identify product features. They assumed that product features are more frequent in product reviews than in a general natural language text. However, their method seems to have low precision since retrieved aspects are affected by noise. Some methods treated the aspect term extraction as sequence labeling and used CRF for that. Such methods have performed very well on the datasets even in cross-domain experiments BIBREF9 , BIBREF10 . Topic modeling has been widely used as a basis to perform extraction and grouping of aspects BIBREF35 , BIBREF36 . Two models were considered: pLSA BIBREF37 and LDA BIBREF38 . Both models introduce a latent variable “topic” between the observable variables “document” and “word” to analyze the semantic topic distribution of documents. In topic models, each document is represented as a random mixture over latent topics, where each topic is characterized by a distribution over words. Such methods have been gaining popularity in social media analysis like emerging political topic detection in Twitter BIBREF39 . The LDA model defines a Dirichlet probabilistic generative process for document-topic distribution; in each document, a latent aspect is chosen according to a multinomial distribution, controlled by a Dirichlet prior INLINEFORM0 . Then, given an aspect, a word is extracted according to another multinomial distribution, controlled by another Dirichlet prior INLINEFORM1 . Among existing works employing these models are the extraction of global aspects ( such as the brand of a product) and local aspects (such as the property of a product BIBREF40 ), the extraction of key phrases BIBREF41 , the rating of multi-aspects BIBREF42 , and the summarization of aspects and sentiments BIBREF43 . BIBREF44 employed the maximum entropy method to train a switch variable based on POS tags of words and used it to separate aspect and sentiment words. BIBREF45 added user feedback to LDA as a response-variable related to each document. BIBREF46 proposed a semi-supervised model. DF-LDA BIBREF47 also represents a semi-supervised model, which allows the user to set must-link and cannot-link constraints. A must-link constraint means that two terms must be in the same topic, while a cannot-link constraint means that two terms cannot be in the same topic. BIBREF48 integrated commonsense in the calculation of word distributions in the LDA algorithm, thus enabling the shift from syntax to semantics in aspect-based sentiment analysis. BIBREF49 proposed two semi-supervised models for product aspect extraction based on the use of seeding aspects. In the category of supervised methods, BIBREF50 employed seed words to guide topic models to learn topics of specific interest to a user, while BIBREF42 and BIBREF51 employed seeding words to extract related product aspects from product reviews. On the other hand, recent approaches using deep CNNs BIBREF52 , BIBREF53 showed significant performance improvement over the state-of-the-art methods on a range of NLP tasks. BIBREF52 fed word embeddings to a CNN to solve standard NLP problems such as named entity recognition (NER), part-of-speech (POS) tagging and semantic role labeling. ### Preliminaries
In this section, we briefly review the theoretical concepts necessary to comprehend the present work. We begin with a description of maximum likelihood estimation of edges in dynamic Gaussian Bayesian networks where each node is a word in a sentence. Next, we show that weights in the CNN can be learned by minimizing a global error function that corresponds to an exponential distribution over a linear combination of input sequence of word features. Notations : Consider a Gaussian network (GN) with time delays which comprises a set of INLINEFORM0 nodes and observations gathered over INLINEFORM1 instances for all the nodes. Nodes can take real values from a multivariate distribution determined by the parent set. Let the dataset of samples be INLINEFORM2 , where INLINEFORM3 represents the sample value of the INLINEFORM4 random variable in instance INLINEFORM5 . Lastly, let INLINEFORM6 be the set of parent variables regulating variable INLINEFORM7 . ### Gaussian Bayesian Networks
In tasks where one is concerned with a specific sentence within the context of the previous discourse, capturing the order of the sequences preceding the one at hand may be particularly crucial. We take as given a sequence of sentences INLINEFORM0 , each in turn being a sequence of words so that INLINEFORM1 , where INLINEFORM2 is the length of sentence INLINEFORM3 . Thus, the probability of a word INLINEFORM4 follows the distribution : DISPLAYFORM0 A Bayesian network is a graphical model that represents a joint multivariate probability distribution for a set of random variables BIBREF54 . It is a directed acyclic graph INLINEFORM0 with a set of parameters INLINEFORM1 that represents the strengths of connections by conditional probabilities. The BN decomposes the likelihood of node expressions into a product of conditional probabilities by assuming independence of non-descendant nodes, given their parents. DISPLAYFORM0 where INLINEFORM0 denotes the conditional probability of node expression INLINEFORM1 given its parent node expressions INLINEFORM2 , and INLINEFORM3 denotes the maximum likelihood(ML) estimate of the conditional probabilities. Figure FIGREF11 (a) illustrates the state space of a Gaussian Bayesian network (GBN) at time instant INLINEFORM0 where each node INLINEFORM1 is a word in the sentence INLINEFORM2 . The connections represent causal dependencies over one or more time instants. The observed state vector of variable INLINEFORM3 is denoted as INLINEFORM4 and the conditional probability of variable INLINEFORM5 given variable INLINEFORM6 is INLINEFORM7 . The optimal Gaussian network INLINEFORM8 is obtained by maximizing the posterior probability of INLINEFORM9 given the data INLINEFORM10 . From Bayes theorem, the optimal Gaussian network INLINEFORM11 is given by: DISPLAYFORM0 where INLINEFORM0 is the probability of the Gaussian network and INLINEFORM1 is the likelihood of the expression data given the Gaussian network. Given the set of conditional distributions with parameters INLINEFORM0 , the likelihood of the data is given by DISPLAYFORM0 To find the likelihood in ( EQREF14 ), and to obtain the optimal Gaussian network as in ( EQREF13 ), Gaussian BN assumes that the nodes are multivariate Gaussian. That is, expression of node INLINEFORM0 can be described with mean INLINEFORM1 and covariance matrix INLINEFORM2 of size INLINEFORM3 . The joint probability of the network can be the product of a set of conditional probability distributions given by: DISPLAYFORM0 where INLINEFORM0 and INLINEFORM1 denotes the regression coefficient matrix, INLINEFORM2 is the conditional variance of INLINEFORM3 given its parent set INLINEFORM4 , INLINEFORM5 is the covariance between observations of INLINEFORM6 and the variables in INLINEFORM7 , and INLINEFORM8 is the covariance matrix of INLINEFORM9 . The acyclic condition of BN does not allow feedback among nodes, and feedback is an essential characteristic of real world GN. Therefore, dynamic Bayesian networks have recently become popular in building GN with time delays mainly due to their ability to model causal interactions as well as feedback regulations BIBREF55 . A first-order dynamic BN is defined by a transition network of interactions between a pair of Gaussian networks connecting nodes at time instants INLINEFORM0 and INLINEFORM1 . In time instant INLINEFORM2 , the parents of nodes are those specified in the time instant INLINEFORM3 . Similarly, the Gaussian network of a INLINEFORM4 -order dynamic system is represented by a Gaussian network comprising INLINEFORM5 consecutive time points and INLINEFORM6 nodes, or a graph of INLINEFORM7 nodes. In practice, the sentence data is transformed to a BOW model where each sentence is a vector of frequencies for each word in the vocabulary. Figure FIGREF11 (b) illustrates the state space of a first-order Dynamic GBN models transition networks among words in sentences INLINEFORM8 and INLINEFORM9 in consecutive time points, the lines correspond to first-order edges among the words learned using BOW. Hence, a sequence of sentences results in a time series of word frequencies. It can be seen that such a discourse model produces compelling discourse vector representations that are sensitive to the structure of the discourse and promise to capture subtle aspects of discourse comprehension, especially when coupled to further semantic data and unsupervised pre-training. ### Convolutional Neural Networks
The idea behind convolution is to take the dot product of a vector of INLINEFORM0 weights INLINEFORM1 also known as kernel vector with each INLINEFORM2 -gram in the sentence INLINEFORM3 to obtain another sequence of features INLINEFORM4 . DISPLAYFORM0 We then apply a max pooling operation over the feature map and take the maximum value INLINEFORM0 as the feature corresponding to this particular kernel vector. Similarly, varying kernel vectors and window sizes are used to obtain multiple features BIBREF23 . For each word INLINEFORM0 in the vocabulary, an INLINEFORM1 dimensional vector representation is given in a look up table that is learned from the data BIBREF30 . The vector representation of a sentence is hence a concatenation of vectors for individual words. Similarly, we can have look up tables for other features. One might want to provide features other than words if these features are suspected to be helpful. Now, the convolution kernels are applied to word vectors instead of individual words. We use these features to train higher layers of the CNN that can represent bigger groups of words in sentences. We denote the feature learned at hidden neuron INLINEFORM0 in layer INLINEFORM1 as INLINEFORM2 . Multiple features may be learned in parallel in the same CNN layer. The features learned in each layer are used to train the next layer DISPLAYFORM0 where * indicates convolution and INLINEFORM0 is a weight kernel for hidden neuron INLINEFORM1 and INLINEFORM2 is the total number of hidden neurons. Training a CNN becomes difficult as the number of layers increases, as the Hessian matrix of second-order derivatives often does not exist. Recently, deep learning has been used to improve the scalability of a model that has inherent parallel computation. This is because hierarchies of modules can provide a compact representation in the form of input-output pairs. Each layer tries to minimize the error between the original state of the input nodes and the state of the input nodes predicted by the hidden neurons. This results in a downward coupling between modules. The more abstract representation at the output of a higher layer module is combined with the less abstract representation at the internal nodes from the module in the layer below. In the next section, we describe deep CNN that can have arbitrary number of layers. ### Convolution Deep Belief Network
A deep belief network (DBN) is a type of deep neural network that can be viewed as a composite of simple, unsupervised models such as restricted Boltzmann machines (RBMs) where each RBMs hidden layer serves as the visible layer for the next RBM BIBREF56 . RBM is a bipartite graph comprising two layers of neurons: a visible and a hidden layer; it is restricted such that the connections among neurons in the same layer are not allowed. To compute the weights INLINEFORM0 of an RBM, we assume that the probability distribution over the input vector INLINEFORM1 is given as: DISPLAYFORM0 where INLINEFORM0 is a normalisation constant. Computing the maximum likelihood is difficult as it involves solving the normalisation constant, which is a sum of an exponential number of terms. The standard approach is to approximate the average over the distribution with an average over a sample from INLINEFORM1 , obtained by Markov chain Monte Carlo until convergence. To train such a multi-layer system, we must compute the gradient of the total energy function INLINEFORM0 with respect to weights in all the layers. To learn these weights and maximize the global energy function, the approximate maximum likelihood contrastive divergence (CD) approach can be used. This method employs each training sample to initialize the visible layer. Next, it uses the Gibbs sampling algorithm to update the hidden layer and then reconstruct the visible layer consecutively, until convergence BIBREF57 . As an example, here we use a logistic regression model to learn the binary hidden neurons and each visible unit is assumed to be a sample from a normal distribution BIBREF58 . The continuous state INLINEFORM0 of the hidden neuron INLINEFORM1 , with bias INLINEFORM2 , is a weighted sum over all continuous visible nodes INLINEFORM3 and is given by: DISPLAYFORM0 where INLINEFORM0 is the connection weight to hidden neuron INLINEFORM1 from visible node INLINEFORM2 . The binary state INLINEFORM3 of the hidden neuron can be defined by a sigmoid activation function: DISPLAYFORM0 Similarly, in the next iteration, the binary state of each visible node is reconstructed and labeled as INLINEFORM0 . Here, we determine the value to the visible node INLINEFORM1 , with bias INLINEFORM2 , as a random sample from the normal distribution where the mean is a weighted sum over all binary hidden neurons and is given by: DISPLAYFORM0 where INLINEFORM0 is the connection weight to hidden neuron INLINEFORM1 from visible node INLINEFORM2 . The continuous state INLINEFORM3 is a random sample from INLINEFORM4 , where INLINEFORM5 is the variance of all visible nodes. Lastly, the weights are updated as the difference between the original and reconstructed visible layer using: DISPLAYFORM0 where INLINEFORM0 is the learning rate and INLINEFORM1 is the expected frequency with which visible unit INLINEFORM2 and hidden unit INLINEFORM3 are active together when the visible vectors are sampled from the training set and the hidden units are determined by ( EQREF21 ). Finally, the energy of a DNN can be determined in the final layer using INLINEFORM4 . To extend the deep belief networks to convolution deep belief network (CDBN) we simply partition the hidden layer into INLINEFORM0 groups. Each of the INLINEFORM1 groups is associated with a INLINEFORM2 filter where INLINEFORM3 is the width of the kernel and INLINEFORM4 is the number of dimensions in the word vector. Let us assume that the input layer has dimension INLINEFORM5 where INLINEFORM6 is the length of the sentence. Then the convolution operation given by ( EQREF17 ) will result in a hidden layer of INLINEFORM7 groups each of dimension INLINEFORM8 . These learned kernel weights are shared among all hidden units in a particular group. The energy function is now a sum over the energy of individual blocks given by: DISPLAYFORM0 The CNN sentence model preserve the order of words by adopting convolution kernels of gradually increasing sizes that span an increasing number of words and ultimately the entire sentence BIBREF31 . However, several word dependencies may occur across sentences hence, in this work we propose a Bayesian CNN model that uses dynamic Bayesian networks to model a sequence of sentences. ### Subjectivity Detection
In this work, we integrate a higher-order GBN for sentences into the first layer of the CNN. The GBN layer of connections INLINEFORM0 is learned using maximum likelihood approach on the BOW model of the training data. The input sequence of sentences INLINEFORM1 are parsed through this layer prior to training the CNN. Only sentences or groups of sentences containing high ML motifs are then used to train the CNN. Hence, motifs are convolved with the input sentences to generate a new set of sentences for pre-training. DISPLAYFORM0 where INLINEFORM0 is the number of high ML motifs and INLINEFORM1 is the training set of sentences in a particular class. Fig. FIGREF28 illustrates the state space of Bayesian CNN where the input layer is pre-trained using a dynamic GBN with up-to two time point delays shown for three sentences in a review on iPhone. The dashed lines correspond to second-order edges among the words learned using BOW. Each hidden layer does convolution followed by pooling across the length of the sentence. To preserve the order of words we adopt kernels of increasing sizes. Since, the number of possible words in the vocabulary is very large, we consider only the top subjectivity clue words to learn the GBN layer. Lastly, In-order to preserve the context of words in conceptual phrases such as `touchscreen'; we consider additional nodes in the Bayesian network for phrases with subjectivity clues. Further, the word embeddings in the CNN are initialized using the log-bilinear language model (LBL) where the INLINEFORM0 dimensional vector representation of each word INLINEFORM1 in ( EQREF10 ) is given by : DISPLAYFORM0 where INLINEFORM0 are the INLINEFORM1 co-occurrence or context matrices computed from the data. The time series of sentences is used to generate a sub-set of sentences containing high ML motifs using ( EQREF27 ). The frequency of a sentence in the new dataset will also correspond to the corresponding number of high ML motifs in the sentence. In this way, we are able to increase the weights of the corresponding causal features among words and concepts extracted using Gaussian Bayesian networks. The new set of sentences is used to pre-train the deep neural network prior to training with the complete dataset. Each sentence can be divided into chunks or phrases using POS taggers. The phrases have hierarchical structures and combine in distinct ways to form sentences. The INLINEFORM0 -gram kernels learned in the first layer hence correspond to a chunk in the sentence. ### Aspect Extraction
In order to train the CNN for aspect extraction, instead, we used a special training algorithm suitable for sequential data, proposed by BIBREF52 . We will summarize it here, mainly following BIBREF59 . The algorithm trains the neural network by back-propagation in order to maximize the likelihood over training sentences. Consider the network parameter INLINEFORM0 . We say that INLINEFORM1 is the output score for the likelihood of an input INLINEFORM2 to have the tag INLINEFORM3 . Then, the probability to assign the label INLINEFORM4 to INLINEFORM5 is calculated as DISPLAYFORM0 Define the logadd operation as DISPLAYFORM0 then for a training example, the log-likelihood becomes DISPLAYFORM0 In aspect term extraction, the terms can be organized as chunks and are also often surrounded by opinion terms. Hence, it is important to consider sentence structure on a whole in order to obtain additional clues. Let it be given that there are INLINEFORM0 tokens in a sentence and INLINEFORM1 is the tag sequence while INLINEFORM2 is the network score for the INLINEFORM3 -th tag having INLINEFORM4 -th tag. We introduce INLINEFORM5 transition score from moving tag INLINEFORM6 to tag INLINEFORM7 . Then, the score tag for the sentence INLINEFORM8 to have the tag path INLINEFORM9 is defined by: DISPLAYFORM0 This formula represents the tag path probability over all possible paths. Now, from ( EQREF32 ) we can write the log-likelihood DISPLAYFORM0 The number of tag paths has exponential growth. However, using dynamic programming techniques, one can compute in polynomial time the score for all paths that end in a given tag BIBREF52 . Let INLINEFORM0 denote all paths that end with the tag INLINEFORM1 at the token INLINEFORM2 . Then, using recursion, we obtain DISPLAYFORM0 For the sake of brevity, we shall not delve into details of the recursive procedure, which can be found in BIBREF52 . The next equation gives the log-add for all the paths to the token INLINEFORM0 : DISPLAYFORM0 Using these equations, we can maximize the likelihood of ( EQREF35 ) over all training pairs. For inference, we need to find the best tag path using the Viterbi algorithm; e.g., we need to find the best tag path that minimizes the sentence score ( EQREF34 ). The features of an aspect term depend on its surrounding words. Thus, we used a window of 5 words around each word in a sentence, i.e., INLINEFORM0 words. We formed the local features of that window and considered them to be features of the middle word. Then, the feature vector was fed to a CNN. The network contained one input layer, two convolution layers, two max-pool layers, and a fully connected layer with softmax output. The first convolution layer consisted of 100 feature maps with filter size 2. The second convolution layer had 50 feature maps with filter size 3. The stride in each convolution layer is 1 as we wanted to tag each word. A max-pooling layer followed each convolution layer. The pool size we use in the max-pool layers was 2. We used regularization with dropout on the penultimate layer with a constraint on L2-norms of the weight vectors, with 30 epochs. The output of each convolution layer was computed using a non-linear function; in our case we used INLINEFORM0 . As features, we used word embeddings trained on two different corpora. We also used some additional features and rules to boost the accuracy; see Section UID49 . The CNN produces local features around each word in a sentence and then combines these features into a global feature vector. Since the kernel size for the two convolution layers was different, the dimensionality INLINEFORM0 mentioned in Section SECREF16 was INLINEFORM1 and INLINEFORM2 , respectively. The input layer was INLINEFORM3 , where 65 was the maximum number of words in a sentence, and 300 the dimensionality of the word embeddings used, per each word. The process was performed for each word in a sentence. Unlike traditional max-likelihood leaning scheme, we trained the system using propagation after convolving all tokens in the sentence. Namely, we stored the weights, biases, and features for each token after convolution and only back-propagated the error in order to correct them once all tokens were processed using the training scheme as explained in Section SECREF30 . If a training instance INLINEFORM0 had INLINEFORM1 words, then we represented the input vector for that instance as INLINEFORM2 . Here, INLINEFORM3 is a INLINEFORM4 -dimensional feature vector for the word INLINEFORM5 . We found that this network architecture produced good results on both of our benchmark datasets. Adding extra layers or changing the pooling size and window size did not contribute to the accuracy much, and instead, only served to increase computational cost. In this subsection, we present the data used in our experiments. BIBREF64 presented two different neural network models for creating word embeddings. The models were log-linear in nature, trained on large corpora. One of them is a bag-of-words based model called CBOW; it uses word context in order to obtain the embeddings. The other one is called skip-gram model; it predicts the word embeddings of surrounding words given the current word. Those authors made a dataset called word2vec publicly available. These 300-dimensional vectors were trained on a 100-billion-word corpus from Google News using the CBOW architecture. We trained the CBOW architecture proposed by BIBREF64 on a large Amazon product review dataset developed by BIBREF65 . This dataset consists of 34,686,770 reviews (4.7 billion words) of 2,441,053 Amazon products from June 1995 to March 2013. We kept the word embeddings 300-dimensional (http://sentic.net/AmazonWE.zip). Due to the nature of the text used to train this model, this includes opinionated/affective information, which is not present in ordinary texts such as the Google News corpus. For training and evaluation of the proposed approach, we used two corpora: Aspect-based sentiment analysis dataset developed by BIBREF66 ; and SemEval 2014 dataset. The dataset consists of training and test sets from two domains, Laptop and Restaurant; see Table TABREF52 . The annotations in both corpora were encoded according to IOB2, a widely used coding scheme for representing sequences. In this encoding, the first word of each chunk starts with a “B-Type” tag, “I-Type” is the continuation of the chunk and “O” is used to tag a word which is out of the chunk. In our case, we are interested to determine whether a word or chunk is an aspect, so we only have “B–A”, “I–A” and “O” tags for the words. Here is an example of IOB2 tags: also/O excellent/O operating/B-A system/I-A ,/O size/B-A and/O weight/B-A for/O optimal/O mobility/B-A excellent/O durability/B-A of/O the/O battery/B-A the/O functions/O provided/O by/O the/O trackpad/B-A is/O unmatched/O by/O any/O other/O brand/O In this section, we present the features, the representation of the text, and linguistic rules used in our experiments. We used the following the features: Word Embeddings We used the word embeddings described earlier as features for the network. This way, each word was encoded as 300-dimensional vector, which was fed to the network. Part of speech tags Most of the aspect terms are either nouns or noun chunk. This justifies the importance of POS features. We used the POS tag of the word as its additional feature. We used 6 basic parts of speech (noun, verb, adjective, adverb, preposition, conjunction) encoded as a 6- dimensional binary vector. We used Stanford Tagger as a POS tagger. These two features vectors were concatenated and fed to CNN. So, for each word the final feature vector is 306 dimensional. In some of our experiments, we used a set of linguistic patterns (LPs) derived from sentic patterns (LP) BIBREF11 , a linguistic framework based on SenticNet BIBREF22 . SenticNet is a concept-level knowledge base for sentiment analysis built by means of sentic computing BIBREF67 , a multi-disciplinary approach to natural language processing and understanding at the crossroads between affective computing, information extraction, and commonsense reasoning, which exploits both computer and human sciences to better interpret and process social information on the Web. In particular, we used the following linguistic rules: Let a noun h be a subject of a word t, which has an adverbial or adjective modifier present in a large sentiment lexicon, SenticNet. Then mark h as an aspect. Except when the sentence has an auxiliary verb, such as is, was, would, should, could, etc., we apply: If the verb t is modified by an adjective or adverb or is in adverbial clause modifier relation with another token, then mark h as an aspect. E.g., in “The battery lasts little”, battery is the subject of lasts, which is modified by an adjective modifier little, so battery is marked as an aspect. If t has a direct object, a noun n, not found in SenticNet, then mark n an aspect, as, e.g., in “I like the lens of this camera”. If a noun h is a complement of a couplar verb, then mark h as an explicit aspect. E.g., in “The camera is nice”, camera is marked as an aspect. If a term marked as an aspect by the CNN or the other rules is in a noun-noun compound relationship with another word, then instead form one aspect term composed of both of them. E.g., if in “battery life”, “battery” or “life” is marked as an aspect, then the whole expression is marked as an aspect. The above rules 1–4 improve recall by discovering more aspect terms. However, to improve precision, we apply some heuristics: e.g., we remove stop-words such as of, the, a, etc., even if they were marked as aspect terms by the CNN or the other rules. We used the Stanford parser to determine syntactic relations in the sentences. We combined LPs with the CNN as follows: both LPs and CNN-based classifier are run on the text; then all terms marked by any of the two classifiers are reported as aspect terms, except for those unmarked by the last rule. Table TABREF63 shows the accuracy of our aspect term extraction framework in laptop and restaurant domains. The framework gave better accuracy on restaurant domain reviews, because of the lower variety of aspect available terms than in laptop domain. However, in both cases recall was lower than precision. Table TABREF63 shows improvement in terms of both precision and recall when the POS feature is used. Pre-trained word embeddings performed better than randomized features (each word's vector initialized randomly); see Table TABREF62 . Amazon embeddings performed better than Google word2vec embeddings. This supports our claim that the former contains opinion-specific information which helped it to outperform the accuracy of Google embeddings trained on more formal text—the Google news corpus. Because of this, in the sequel we only show the performance using Amazon embeddings, which we denote simply as WE (word embeddings). In both domains, CNN suffered from low recall, i.e., it missed some valid aspect terms. Linguistic analysis of the syntactic structure of the sentences substantially helped to overcome some drawbacks of machine learning-based analysis. Our experiments showed good improvement in both precision and recall when LPs were used together with CNN; see Table TABREF64 . As to the LPs, the removal of stop-words, Rule 1, and Rule 3 were most beneficial. Figure FIGREF66 shows a visualization for the Table TABREF64 . Table TABREF65 and Figure FIGREF61 shows the comparison between the proposed method and the state of the art on the Semeval dataset. It is noted that about 36.55% aspect terms present in the laptop domain corpus are phrase and restaurant corpus consists of 24.56% aspect terms. The performance of detecting aspect phrases are lower than single word aspect tokens in both domains. This shows that the sequential tagging is indeed a tough task to do. Lack of sufficient training data for aspect phrases is also one of the reasons to get lower accuracy in this case. In particular, we got 79.20% and 83.55% F-score to detect aspect phrases in laptop and restaurant domain respectively. We observed some cases where only 1 term in an aspect phrase is detected as aspect term. In those cases Rule 4 of the LPs helped to correctly detect the aspect phrases. We also carried out experiments on the aspect dataset originally developed by BIBREF66 . This is to date the largest comprehensive aspect-based sentiment analysis dataset. The best accuracy on this dataset was obtained when word embedding features were used together with the POS features. This shows that while the word embedding features are most useful, the POS feature also plays a major role in aspect extraction. As on the SemEval dataset, LPs together with CNN increased the overall accuracy. However, LPs have performed much better on this dataset than on the SemEval dataset. This supports the observation made previously BIBREF66 that on this dataset LPs are more useful. One of the possible reasons for this is that most of the sentences in this dataset are grammatically correct and contain only one aspect term. Here we combined LPs and a CNN to achieve even better results than the approach of by BIBREF66 based only on LPs. Our experimental results showed that this ensemble algorithm (CNN+LP) can better understand the semantics of the text than BIBREF66 's pure LP-based algorithm, and thus extracts more salient aspect terms. Table TABREF69 and Figure FIGREF68 shows the performance and comparisons of different frameworks. Figure FIGREF70 compares the proposed method with the state of the art. We believe that there are two key reasons for our framework to outperform state-of-the-art approaches. First, a deep CNN, which is non-linear in nature, better fits the data than linear models such as CRF. Second, the pre-trained word embedding features help our framework to outperform state-of-the-art methods that do not use word embeddings. The main advantage of our framework is that it does not need any feature engineering. This minimizes development cost and time. ### Subjectivity Detection
We use the MPQA corpus BIBREF20 , a collection of 535 English news articles from a variety of sources manually annotated with subjectivity flag. From the total of 9,700 sentences in this corpus, 55 INLINEFORM0 of the sentences are labeled as subjective while the rest are objective. We also compare with the Movie Review (MR) benchmark dataset BIBREF28 , that contains 5000 subjective movie review snippets from Rotten Tomatoes website and another 5000 objective sentences from plot summaries available from the Internet Movies Database. All sentences are at least ten words long and drawn from reviews or plot summaries of movies released post 2001. The data pre-processing included removing top 50 stop words and punctuation marks from the sentences. Next, we used a POS tagger to determine the part-of-speech for each word in a sentence. Subjectivity clues dataset BIBREF19 contains a list of over 8,000 clues identified manually as well as automatically using both annotated and unannotated data. Each clue is a word and the corresponding part of speech. The frequency of each clue was computed in both subjective and objective sentences of the MPQA corpus. Here we consider the top 50 clue words with highest frequency of occurrence in the subjective sentences. We also extracted 25 top concepts containing the top clue words using the method described in BIBREF11 . The CNN is collectively pre-trained with both subjective and objective sentences that contain high ML word and concept motifs. The word vectors are initialized using the LBL model and a context window of size 5 and 30 features. Each sentence is wrapped to a window of 50 words to reduce the number of parameters and hence the over-fitting of the model. A CNN with three hidden layers of 100 neurons and kernels of size INLINEFORM0 is used. The output layer corresponds to two neurons for each class of sentiments. We used 10 fold cross validation to determine the accuracy of classifying new sentences using the trained CNN classifier. A comparison is done with classifying the time series data using baseline classifiers such as Naive Bayes SVM (NBSVM) BIBREF60 , Multichannel CNN (CNN-MC) BIBREF61 , Subjectivity Word Sense Disambiguation (SWSD) BIBREF62 and Unsupervised-WSD (UWSD) BIBREF63 . Table TABREF41 shows that BCDBN outperforms previous methods by INLINEFORM0 in accuracy on both datasets. Almost INLINEFORM1 improvement is observed over NBSVM on the movie review dataset. In addition, we only consider word vectors of 30 features instead of the 300 features used by CNN-MC and hence are 10 times faster. ### Key Applications
Subjectivity detection can prevent the sentiment classifier from considering irrelevant or potentially misleading text. This is particularly useful in multi-perspective question answering summarization systems that need to summarize different opinions and perspectives and present multiple answers to the user based on opinions derived from different sources. It is also useful to analysts in government, commercial and political domains who need to determine the response of the people to different crisis events. After filtering of subjective sentences, aspect mining can be used to provide clearer visibility into the emotions of people by connecting different polarities to the corresponding target attribute. ### Conclusion
In this chapter, we tackled the two basic tasks of sentiment analysis in social media: subjectivity detection and aspect extraction. We used an ensemble of deep learning and linguistics to collect opinionated information and, hence, perform fine-grained (aspect-based) sentiment analysis. In particular, we proposed a Bayesian deep convolutional belief network to classify a sequence of sentences as either subjective or objective and used a convolutional neural network for aspect extraction. Coupled with some linguistic rules, this ensemble approach gave a significant improvement in performance over state-of-the-art techniques and paved the way for a more multifaceted (i.e., covering more NLP subtasks) and multidisciplinary (i.e., integrating techniques from linguistics and other disciplines) approach to the complex problem of sentiment analysis. ### Future Directions
In the future we will try to visualize the hierarchies of features learned via deep learning. We can also consider fusion with other modalities such as YouTube videos. ### Acknowledgement
This work was funded by Complexity Institute, Nanyang Technological University. ### Cross References
Sentiment Quantification of User-Generated Content, 110170 Semantic Sentiment Analysis of Twitter Data, 110167 Twitter Microblog Sentiment Analysis, 265 Fig. 1 State space of different Bayesian models Fig. 2 State space of Bayesian CNN where the input layer is pre-trained using a dynamic GBN Table 2 SemEval Data used for Evaluation Fig. 3 Comparison of the performance with the state of the art. Table 3 Random features vs. Google Embeddings vs. Amazon Embeddings on the SemEval 2014 dataset Table 4 Feature analysis for the CNN classifier Table 5 Impact of Sentic Patterns on the SemEval 2014 dataset Table 6 Comparison with the state of the art. ZW stands for [68]; LP stands for Sentic Patterns. Fig. 4 Comparison between the performance of CNN, CNN-LP and LP. Table 7 Impact of the POS feature on the dataset by [52] Fig. 5 Comparison between the performance of CNN, CNN-LP and LP. Table 8 Impact of Sentic Patterns on the dataset by [52] Fig. 6 Comparison of the performance with the state of the art on Bing Liu dataset. | apply an ensemble of deep learning and linguistics t |
Where do the concept sets come from? | ### Introduction
Commonsense reasoning has long been acknowledged as a critical bottleneck of artificial intelligence and especially in natural language processing. It is an ability of combining commonsense facts and logic rules to make new presumptions about ordinary scenes in our daily life. A distinct property of commonsense reasoning problems is that they are generally trivial for human-beings while challenging for machine reasoners. There have been a few recent tasks and datasets for testing machine commonsense, while most of them frame their problems as multi-choice question answering, such as CSQA BIBREF0 and SWAG BIBREF1. We name this kind of tasks as deterministic commonsense reasoning because they focus on modeling the plausibility of given complete scenes. The systems for these tasks thus have to work with biased selection of distractors, and thus are less practical or challenging. Simply fine-tuning such large pre-trained language encoders can yield near or exceeding human performance BIBREF2. On the other hand, few work has been done so far in testing machine commonsense in a generative reasoning setting, where a reasoner is expected to complete scenes with several given concepts. Specifically, we would like to investigate if machine-reasoning models can generate a sentence that contains a required set of concepts (i.e. nouns or verbs) while describing a common scene in our daily life. For example, as shown in Figure FIGREF1, given an unordered collection of concepts “{apple (noun), bag (noun), pick (verb), place (verb), tree (noun)}”, a rational reasoner should be able to generate a sentence like “A boy picks some apples from a tree and places them into a bag.”, which describes a natural scene and contains all given concepts. The creation of this sentence is easy for humans while non-trivial for even state-of-the-art conditional language generation models. We argue that such an ability of recovering natural scenes of daily life can benefit a wide range of natural language generation (NLG) tasks including image/video captioning BIBREF3, BIBREF4, scene-based visual reasoning and VQA BIBREF5, storytelling BIBREF6, and dialogue systems BIBREF7, BIBREF8. Towards empowering machines with the generative commonsense reasoning ability, we create a large-scale dataset, named CommonGen, for the constrained text generation task. We collect $37,263$ concept-sets as the inputs, each of which contains three to five common concepts. These concept-sets are sampled from several large corpora of image/video captions, such that the concepts inside them are more likely to co-occur in natural scenes. Through crowd-sourcing via Amazon Mechanical Turk (AMT), we finally obtain $89,028$ human-written sentences as expected outputs. We investigate the performance of sophisticated sequence generation methods for the proposed task with both automatic metrics and human evaluation. The experiments show that all methods are far from human performance in generative commonsense reasoning. Our main contributions are as follows: 1) We introduce the first large-scale constrained text generation dataset targeting at generative commonsense reasoning; 2) We systematically compare methods for this (lexically) constrained text generation with extensive experiments and evaluation. 3) Our code and data are publicly available (w/ the URL in the abstract), so future research in this direction can be directly developed in a unified framework. ### Problem Formulation
In this section, we formulate our task with mathematical notations and discuss its inherent challenges. The input to the task is a set of $n$ concepts $x=\lbrace c_1,c_2,\dots ,c_n\rbrace \in \mathcal {X}$, where $c_i\in \mathcal {C}$ is a common noun or verb. $\mathcal {X}$ denotes the space of concept-sets and $\mathcal {C}$ stands for the concept vocabulary. The expected output of this task is a simple, grammatical sentence $y\in \mathcal {Y}$, describing a natural scene in our daily-life that covers all given concepts in $x$. Note that other forms of given concepts are also accepted, such as plural forms of nouns and verbs. In addition, we also provide rationales as an optional resource to model the generation process. For each pair of $(x, y)$, a rationale $r$ is a list of sentences that explains the background commonsense knowledge used in the scene recovering process. The task is to learn a structured predictive function $f:\mathcal {X} \rightarrow \mathcal {Y}$, which maps a concept-set to a sentence. Thus, it can be seen as a special case of constrained text generation BIBREF9. The unique challenges of our proposed task come from two main aspects as follows. Constrained Decoding. Lexically constrained decoding for sentence generation has been an important and challenging research topic in machine translation community BIBREF10, where they focus on how to decode sentences when some words/phrases (e.g. terminology) must present in target sentences (Section SECREF6). However, it is still an open problem how to efficiently generate sentences given an unordered set of multiple keywords with potential morphological changes (e.g. “pick” $\rightarrow $ “picks” in the previous case). Apart from that, the part-of-speech constraints brings even more difficulties (e.g. “place” can be verb/noun). Commonsense Reasoning. Apart from the challenge in constrained decoding, a generative commonsense reasoner also has to compositionally use (latent) commonsense knowledge for generating most plausible scenes. Recall the illustrative example in Figure FIGREF1, even such a simple scene generation process needs pretty much commonsense knowledge like: 1) “apples grow in trees”; 2) “bags are containers that you can put something in”; 3) “you usually pick something and then place it in a container”. Expected reasoners have to prioritize target scenes over an infinity number of less plausible scenes like “A boy picks an apple tree and places it into bags.” or “A boy places some bags on a tree and picks an apple.”. ### The CommonGen Dataset
In this section, we present how we build the CommonGen dataset for testing machine commonsense with generative reasoning. The overall data collection process is as follows. 1) We first collect a large amount of high-quality image/video caption sentences from several existing corpora, 2) Then, we compute co-occurrence statistics about concept-sets of different sizes ($3\sim 5$), such that we can find the concept-sets that are more likely to be present in the same scene. 3) Finally, we ask human crowd-workers from AMT to write scenes with rationales for every given concept-set, which serve as our development and test sets. The training set consists of carefully post-processed human-written caption sentences, which have little overlap with dev/test sets. We present the statistics and show its inherent challenges at the end of this section. ### The CommonGen Dataset ::: Collecting Concept-Sets with Captions
Following the general definition in the largest commonsense knowledge graph, ConceptNet BIBREF11, we understand a concept as a common noun or verb. We aim to test the ability of generating natural scenes with a given set of concepts. The expected concept-sets in our task are supposed to be likely co-occur in natural, daily-life scenes . The concepts in images/videos captions, which usually describe scenes in our daily life, thus possess the desired property. We therefore collect a large amount of caption sentences from a variety of datasets, including VATEX BIBREF4, LSMDC BIBREF12, ActivityNet BIBREF13, and SNLI BIBREF15, forming 1,040,330 sentences in total. We assume if a set of concepts are all mentioned together in more caption sentences, then this concept-set is more like to co-occur. Thus, we compute the co-occurrence frequency of all possible concept-sets that have $3\sim 5$ concepts, named as three/four/five-concept-sets respectively. Each concept-set is associated with at least one caption sentences. We carefully post-process them and take the shortest ones with minimal overlaps as the final data. These initial concept-sets are further divided into three parts: train/dev/test. We then iterate all training concept-sets and remove the ones that have more than two overlapping concepts with any concept-set in the dev or test set. Thus, the dev/test set can better measure the generalization ability of models on unseen combinations of concepts. ### The CommonGen Dataset ::: Crowd-Sourcing via AMT
It is true that the above-mentioned associated caption sentences for each concept-set are human-written and do describe scenes that cover all given concepts. However, they are created under specific contexts (i.e. an image or a video) and thus might be less representative for common sense. To better measure the quality and interpretability of generative reasoners, we need to evaluate them with scenes and rationales created by using concept-sets only as the signals for annotators. We collect more human-written scenes for each concept-set in dev and test set through crowd-sourcing via the Amazon Mechanical Turk platform. Each input concept-set is annotated by at least three different humans. The annotators are also required to give sentences as the rationales, which further encourage them to use common sense in creating their scenes. The crowd-sourced sentences correlate well with the associated captions, meaning that it is reasonable to use caption sentences as training data although they can be partly noisy. Additionally, we utilize a search engine over the OMCS corpus BIBREF16 for retrieving relevant propositions as distant rationales in training data. ### The CommonGen Dataset ::: Statistics
We present the statistical information of our final dataset. Firstly, we summarize the basic statistics in Table TABREF9, such as the number of unique concept-sets, scene sentences, and sentence lengths. In total, there are 3,706 unique concepts among all concept-sets, and 3,614/1,018/1,207 in the train/dev/test parts respectively. Note that there are 4% of the dev and 6% of the test concepts never appear in the training data, so we can better understand how well trained models can perform with unseen concepts. We analyze the overlap between training concept-sets and dev/test concept-sets. By average, we find that 98.8% of the training instances share no common concept at all with dev/test data, such that the dev/test can help us analyze model performance on new combinations of concepts. We also visualize the frequency distribution of our test concept-sets in Figure FIGREF7 by showing the frequency of top 50 single concepts and co-occurred concept pairs. ### Methods
In this section, we introduce the methods that we adopt for the proposed constrained text generation task. We group these methods into several types as follows. Basically, we have different kinds of encoder-decoder architectures with copy attention mechanism, including both classic and recently proposed methods. Apart from that, we utilize the state-of-the-art pre-trained sentence generation model for our task. Moreover, we include three typical models for abstractive summarization, story generation respectively, and keywords-based decoding of language models. ### Methods ::: Seq-to-Seq Learning
One very straightforward way is to form this problem as a “sequence”-to-sequence task, where input sequences are randomly sorted sets of given concepts. In this way, encoder-decoder seq2seq architectures based on bidirectional RNN (bRNN) BIBREF17 or Transformer (Trans.) BIBREF18 can be directly adopted to the task, just like many other conditional sequence generation problems (translation, summarization, etc.). Order-insensitive processing. However, these encoders may degrade because our inputs are actually order-insensitive. We thus try to use multi-layer perceptrons (MLP) with mean-pooling as the encoder (“mean encoder”) over sequences of word vectors to completely eliminate the order sensitivity. Similarly, we consider removing the positional embeddings in Transformers (Trans. w/o Pos). Copying mechanism. The above-mentioned architectures with vanilla attention can miss the words in input sequences and thus produce either unknown tokens or synonyms. To force the decoder to produce target sentences with a constraint on input sentence, we utilize the copying mechanism BIBREF19 for all these models. We follow the implementation of these methods by OpenNMT-py BIBREF20. Non-autoregressive generation. Recent advances in conditional sentence generation have a focus on edit-based models, which iteratively refine generated sequences (usually bounded by a fixed length). These models potentially get better performance than auto-regressive methods because of their explicit modeling on iterative refinements. We study typical models including iNAT BIBREF21, Insertion Transformer (InsertTrans) BIBREF22, and Levenshtein Transformer (LevenTrans) BIBREF23. ### Methods ::: A BERT-based Method: UniLM
We employ a new unified pre-trained language model, UniLM BIBREF24, which uses BERT BIBREF25 as the encoder and then fine-tunes the whole architecture with different generation-based objective. To the best of our knowledge, the UniLM model is the state-of-the-art method for a wide range of conditional text generation tasks including summarization, question generation, and dialogue responding. ### Methods ::: Other methods
Based on the similarity between our task and abstractive summarization and story generation (with given topic words), we also apply Pointer Generator Networks (“PointerGen”) BIBREF26 and Multi-scale Fusion Attention (“Fusion Attn.”) BIBREF27 model respectively for our task. ### Methods ::: Incorporating Commonsense Rationales
We explore how to utilize additional commonsense knowledge (i.e. rationales) as the input to the task. Like we mentioned in Section SECREF6, we search relevant sentences from the OMCS corpus as the additional distant rationales, and ground truth rationale sentences for dev/test data. The inputs are no longer the concept-sets themselves, but in a form of “[rationales$|$concept-set]” (i.e. concatenating the rationale sentences and original concept-set strings). ### Evaluation
Herein, we present the experimental results for comparing different baseline methods in the proposed setting. We first introduce the setup and automatic metrics, and then we present the results and analysis. Finally, we show human evaluation results and qualitative analysis. ### Evaluation ::: Setup
We use the proposed CommonGen dataset in two setting: knowledge-agnostic and knowledge-aware. For the knowledge-agnostic setting, we simply apply the methods in Section SECREF4 while we concatenate rationales and input concept-sets together as the knowledge-aware inputs (“$+r$”). ### Evaluation ::: Automatic Metrics
For automatically evaluating our methods, we propose to use widely used metric for image/video captioning. This is because the proposed CommonGen task can be regarded as also a caption task where the context are incomplete scenes with given concept-sets. Therefore, we choose BLEU-3/4 BIBREF28, ROUGE-2/L BIBREF29, CIDEr BIBREF30, and SPICE BIBREF31 as the main metrics. Apart from these classic metrics, we also include a novel embedding-based metric named BERTScore BIBREF32. To make the comparisons more clear, we show the delta of BERTScore results by subtracting the score of merely using input concept-sets as target sentences, named $\triangle $BERTS. To have an estimation about human performance in each metric, we iteratively treat every reference sentence in dev/test data as the prediction to be compared with all references (including itself). That is, if a model has the same reasoning ability with average performance of our crowd workers, its results should exceed this “human bound”. ### Evaluation ::: Experimental Results
We present the experimental results of five groups of methods that are introduced in Section SECREF4. We find that the model UniLM outperforms all other baseline methods by a large margin, which is expected due to it is pre-trained with the BERT encoder towards generation objectives. However, its performance is still way far from the human bound, and this margin is even larger in test data. We notice that the most recent edit-based model named LevenTrans archives the best performance among models without pre-training at all. This shows that edit-based sequence generation models can better deal with the cases where target sentences share similar vocabulary with source ones. Nonetheless, the other two models within the same sequence modeling framework (i.e. fairseq) are much worse, which might because of their specialty designed for machine translation. An order-insensitive sequence/set encoder, “mean encoder”, outperform order-sensitive counterparts like “bRNN”. However, such a marginal improvement is not seen in the comparison between “Trans.” vs “Trans. w/o Pos”. We assume that for short sequences the order sensitivity does not harm sequential encoders, while positional embeddings in Transformers can better improve the self-attention mechanism. Also, we find that Transformer-based seq2seq architectures are not outperforming simpler models like bRNN. As for the use of additional retrieved sentences form OMCS corpus and human-written associated rationales, we find that they are not generally helpful in investigated architectures. Although they increase the BLEU and ROUGE scores, the metrics specially designed for captioning like CIDEr and SPICE are dropping down. We argue that it might because the OMCS sentences are actually not aligned with training data, and more sophisticated methods for encoding such non-sequential facts in a more compositional way. ### Evaluation ::: Human Evaluation
From the automatic evaluation results with multiple metrics, we have a rough idea of the performance of all models. However, no automatic metric is perfect, especially for a newly proposed generation task like the CommonGen. We thus ask humans to rank 100 outputs of 6 selected typical models as well as one randomly picked reference sentence, forming seven systems in total. Annotators are educated to rank results by their coverage, fluency, and plausibility in daily life. Then, we compute the cumulative gains of each system in all 100 cases: $S^{(k)}_i$ is the final score of the $i$-th system by the $k$-th annotator. $G^{k}_{i, j}$ is the rank position of the $i$-th system output for $j$-th example. In our case, $N=100$, $K = 5$, $G^{k}_{i, j}\in [1,7]$. As shown in Table TABREF22, we compare different systems including human bound for both the above-introduced cumulative ranking scores and the average hit@top3 rates with standard deviations. We find that the correlation between human evaluation and CIDEr and SPICE are better than the other metrics (see Table TABREF15). ### Evaluation ::: Qualitative Analysis
For more clearly observe the performance of interested models, we present several real system outputs on the test set in Table TABREF24. We find that models usually cannot cover all given concepts, and also can produce repetitions of given concepts (e.g. “a dog catches a dog”, “a couple of couples”, and “at an object and an object .”). Moreover, we find that the order of actions may be mot natural. For example, the model output “a man pulls a sword out of his mouth and swallows it” makes less sense because a man usually swallow a sword first before he pull it out in such performances. ### Related Work ::: Machine Common Sense
Machine common sense (MCS) has long been considered as one of the most significant area in artificial intelligence. Recently, there are various emerging datasets for testing machine commonsense from different angles, such as commonsense extraction BIBREF33, BIBREF34, next situation prediction (SWAG BIBREF1, CODAH BIBREF35, HellaSWAG BIBREF36), cultural/social understanding BIBREF37, BIBREF38, BIBREF39, visual scene comprehension BIBREF40, and general commonsense question answering BIBREF0, BIBREF41. Most of them are in a multi-choice QA setting for discriminative commonsense reasoning, among which CSQA BIBREF0 and SWAG BIBREF1 are two typical examples. The input of the CSQA task is a question that needs commonsense reasoning and there are five candidate answers (words/phrases). The SWAG task asks models to select which situation is the most plausible next situation, given a sentence describing an event. The two tasks share very similar objectives with large pre-trained language encoders like BERT BIBREF42: Masked-LM can predict the missing words in an incomplete sentence, which is similar to the CSQA setting; NextSentPrediction classifies whether a sentence is the next sentence of the given sentence in the corpora, which can be seen as using distant supervision for the SWAG task. Thus, simply fine-tuning such large pre-trained language encoders can yield near or exceeding human performance BIBREF43, BIBREF2, but it does not necessarily mean machine reasoners can really produce new assumptions in an open and generative setting. The proposed CommonGen, to the best of our knowledge, is the first dataset and task for generative commonsense reasoning. ### Related Work ::: Constrained Text Generation
Constrained or controllable text generation aims to decode realistic sentences that have expected attributes such as sentiment BIBREF44, BIBREF9, tense BIBREF9, template BIBREF45, style BIBREF46, BIBREF47, BIBREF48, etc. The most similar scenario with our task is lexically constrained sentence encoding, which has been studied mainly in the machine translation community BIBREF49, BIBREF50 for dealing with terminology and additional bilingual dictionaries. Classic methods usually modify the (beam) searching algorithms to accommodate lexical constraints like Grid Beam Search BIBREF10. The most recent work in this line is the CGMH BIBREF51 model, which works in the inference stage to sample sentences with a sequence of multiple keywords from language models. However, our task brings more challenges: 1) we do not assume there is a fixed order of keywords in target sentences; 2) we allow morphological changes of the keywords; 3) the decoded sentences must describe highly plausible scenes in our daily life. Current methods cannot well address these issues and also work extremely slow to generate grammatical sentences. We instead mainly investigate sequence-to-sequence architectures, especially models that are based on editing operations and non-autoregressive. Pre-trained seq2seq generation models like UniLM BIBREF24 and BRAT BIBREF52 are usually initialized with pre-trained language encoder and then further fine-tuned with multiple NLG tasks. The UniLM archives the best performance on our proposed CommonGen task, while being far from human-level performance and hardly interpretable. ### Conclusion
In this paper, we purpose a novel constrained text generation task for generative commonsense reasoning. We introduce a new large-scale dataset named CommonGen and investigate various methods on them. Through our extensive experiments and human evaluation, we demonstrate that the inherent difficulties of the new task cannot be addressed by even the state-of-the-art pre-trained language generation model. For the future research, we believe the following directions are highly valuable to explore: 1) specially designed metrics for automatic evaluation that focus on commonsense plausibility; 2) better mechanisms for retrieving and imposing useful commonsense knowledge into sentence generation processes; 3) explicitly modeling keyword-centric edits (e.g. insertion, deletion, morphological changes) such that relevant commonsense knowledge can be well utilized. We also believe that models performed well on CommonGen can be easily transferred to other commonsense-required reasoning tasks with few annotations, including image/video captioning, visual question answering, and discriminative multi-choice commonsense question answering. Figure 1: A motivating example for generative commonsense reasoning and the COMMONGEN task. A reasoner gets a concept-set as the input and should generate a sentence that covers all given concepts while describing a common scene (in the green box) out of less plausible ones (in the red box). Figure 2: The frequency of top 50 single concepts (upper) and co-occurred concept-pairs (lower) in the test data. Table 1: The basic statistics of COMMONGEN. Table 2: Experimental results of different baseline methods on the COMMONGEN. Table 3: The average humane evaluation ranking scores and hit@top3 rates for each tested system. | These concept-sets are sampled from several large corpora of image/video captions |
What dataset/corpus is this evaluated over? | ### Introduction
Neural models have recently gained popularity for Natural Language Processing (NLP) tasks BIBREF0 , BIBREF1 , BIBREF2 . For sentence classification, in particular, Convolution Neural Networks (CNN) have realized impressive performance BIBREF3 , BIBREF4 . These models operate over word embeddings, i.e., dense, low dimensional vector representations of words that aim to capture salient semantic and syntactic properties BIBREF1 . An important consideration for such models is the specification of the word embeddings. Several options exist. For example, Kalchbrenner et al. kalchbrenner2014convolutional initialize word vectors to random low-dimensional vectors to be fit during training, while Johnson and Zhang johnson2014effective use fixed, one-hot encodings for each word. By contrast, Kim kim2014convolutional initializes word vectors to those estimated via the word2vec model trained on 100 billion words of Google News BIBREF5 ; these are then updated during training. Initializing embeddings to pre-trained word vectors is intuitively appealing because it allows transfer of learned distributional semantics. This has allowed a relatively simple CNN architecture to achieve remarkably strong results. Many pre-trained word embeddings are now readily available on the web, induced using different models, corpora, and processing steps. Different embeddings may encode different aspects of language BIBREF6 , BIBREF7 , BIBREF8 : those based on bag-of-words (BoW) statistics tend to capture associations (doctor and hospital), while embeddings based on dependency-parses encode similarity in terms of use (doctor and surgeon). It is natural to consider how these embeddings might be combined to improve NLP models in general and CNNs in particular. Contributions. We propose MGNC-CNN, a novel, simple, scalable CNN architecture that can accommodate multiple off-the-shelf embeddings of variable sizes. Our model treats different word embeddings as distinct groups, and applies CNNs independently to each, thus generating corresponding feature vectors (one per embedding) which are then concatenated at the classification layer. Inspired by prior work exploiting regularization to encode structure for NLP tasks BIBREF9 , BIBREF10 , we impose different regularization penalties on weights for features generated from the respective word embedding sets. Our approach enjoys the following advantages compared to the only existing comparable model BIBREF11 : (i) It can leverage diverse, readily available word embeddings with different dimensions, thus providing flexibility. (ii) It is comparatively simple, and does not, for example, require mutual learning or pre-training. (iii) It is an order of magnitude more efficient in terms of training time. ### Related Work
Prior work has considered combining latent representations of words that capture syntactic and semantic properties BIBREF12 , and inducing multi-modal embeddings BIBREF13 for general NLP tasks. And recently, Luo et al. luo2014pre proposed a framework that combines multiple word embeddings to measure text similarity, however their focus was not on classification. More similar to our work, Yin and Schütze yin-schutze:2015:CoNLL proposed MVCNN for sentence classification. This CNN-based architecture accepts multiple word embeddings as inputs. These are then treated as separate `channels', analogous to RGB channels in images. Filters consider all channels simultaneously. MVCNN achieved state-of-the-art performance on multiple sentence classification tasks. However, this model has practical drawbacks. (i) MVCNN requires that input word embeddings have the same dimensionality. Thus to incorporate a second set of word vectors trained on a corpus (or using a model) of interest, one needs to either find embeddings that happen to have a set number of dimensions or to estimate embeddings from scratch. (ii) The model is complex, both in terms of implementation and run-time. Indeed, this model requires pre-training and mutual-learning and requires days of training time, whereas the simple architecture we propose requires on the order of an hour (and is easy to implement). ### Model Description
We first review standard one-layer CNN (which exploits a single set of embeddings) for sentence classification BIBREF3 , and then propose our augmentations, which exploit multiple embedding sets. Basic CNN. In this model we first replace each word in a sentence with its vector representation, resulting in a sentence matrix INLINEFORM0 , where INLINEFORM1 is the (zero-padded) sentence length, and INLINEFORM2 is the dimensionality of the embeddings. We apply a convolution operation between linear filters with parameters INLINEFORM3 and the sentence matrix. For each INLINEFORM4 , where INLINEFORM5 denotes `height', we slide filter INLINEFORM6 across INLINEFORM7 , considering `local regions' of INLINEFORM8 adjacent rows at a time. At each local region, we perform element-wise multiplication and then take the element-wise sum between the filter and the (flattened) sub-matrix of INLINEFORM9 , producing a scalar. We do this for each sub-region of INLINEFORM10 that the filter spans, resulting in a feature map vector INLINEFORM11 . We can use multiple filter sizes with different heights, and for each filter size we can have multiple filters. Thus the model comprises INLINEFORM12 weight vectors INLINEFORM13 , each of which is associated with an instantiation of a specific filter size. These in turn generate corresponding feature maps INLINEFORM14 , with dimensions varying with filter size. A 1-max pooling operation is applied to each feature map, extracting the largest number INLINEFORM15 from each feature map INLINEFORM16 . Finally, we combine all INLINEFORM17 together to form a feature vector INLINEFORM18 to be fed through a softmax function for classification. We regularize weights at this level in two ways. (1) Dropout, in which we randomly set elements in INLINEFORM19 to zero during the training phase with probability INLINEFORM20 , and multiply INLINEFORM21 with the parameters trained in INLINEFORM22 at test time. (2) An l2 norm penalty, for which we set a threshold INLINEFORM23 for the l2 norm of INLINEFORM24 during training; if this is exceeded, we rescale the vector accordingly. For more details, see BIBREF4 . MG-CNN. Assuming we have INLINEFORM0 word embeddings with corresponding dimensions INLINEFORM1 , we can simply treat each word embedding independently. In this case, the input to the CNN comprises multiple sentence matrices INLINEFORM2 , where each INLINEFORM3 may have its own width INLINEFORM4 . We then apply different groups of filters INLINEFORM5 independently to each INLINEFORM6 , where INLINEFORM7 denotes the set of filters for INLINEFORM8 . As in basic CNN, INLINEFORM9 may have multiple filter sizes, and multiple filters of each size may be introduced. At the classification layer we then obtain a feature vector INLINEFORM10 for each embedding set, and we can simply concatenate these together to form the final feature vector INLINEFORM11 to feed into the softmax function, where INLINEFORM12 . This representation contains feature vectors generated from all sets of embeddings under consideration. We call this method multiple group CNN (MG-CNN). Here groups refer to the features generated from different embeddings. Note that this differs from `multi-channel' models because at the convolution layer we use different filters on each word embedding matrix independently, whereas in a standard multi-channel approach each filter would consider all channels simultaneously and generate a scalar from all channels at each local region. As above, we impose a max l2 norm constraint on the final feature vector INLINEFORM13 for regularization. Figure FIGREF1 illustrates this approach. MGNC-CNN. We propose an augmentation of MG-CNN, Multi-Group Norm Constraint CNN (MGNC-CNN), which differs in its regularization strategy. Specifically, in this variant we impose grouped regularization constraints, independently regularizing subcomponents INLINEFORM0 derived from the respective embeddings, i.e., we impose separate max norm constraints INLINEFORM1 for each INLINEFORM2 (where INLINEFORM3 again indexes embedding sets); these INLINEFORM4 hyper-parameters are to be tuned on a validation set. Intuitively, this method aims to better capitalize on features derived from word embeddings that capture discriminative properties of text for the task at hand by penalizing larger weight estimates for features derived from less discriminative embeddings. ### Datasets
Stanford Sentiment Treebank Stanford Sentiment Treebank (SST) BIBREF14 . This concerns predicting movie review sentiment. Two datasets are derived from this corpus: (1) SST-1, containing five classes: very negative, negative, neutral, positive, and very positive. (2) SST-2, which has only two classes: negative and positive. For both, we remove phrases of length less than 4 from the training set. Subj BIBREF15 . The aim here is to classify sentences as either subjective or objective. This comprises 5000 instances of each. TREC BIBREF16 . A question classification dataset containing six classes: abbreviation, entity, description, human, location and numeric. There are 5500 training and 500 test instances. Irony BIBREF17 . This dataset contains 16,006 sentences from reddit labeled as ironic (or not). The dataset is imbalanced (relatively few sentences are ironic). Thus before training, we under-sampled negative instances to make classes sizes equal. Note that for this dataset we report the Area Under Curve (AUC), rather than accuracy, because it is imbalanced. ### Pre-trained Word Embeddings
We consider three sets of word embeddings for our experiments: (i) word2vec is trained on 100 billion tokens of Google News dataset; (ii) GloVe BIBREF18 is trained on aggregated global word-word co-occurrence statistics from Common Crawl (840B tokens); and (iii) syntactic word embedding trained on dependency-parsed corpora. These three embedding sets happen to all be 300-dimensional, but our model could accommodate arbitrary and variable sizes. We pre-trained our own syntactic embeddings following BIBREF8 . We parsed the ukWaC corpus BIBREF19 using the Stanford Dependency Parser v3.5.2 with Stanford Dependencies BIBREF20 and extracted (word, relation+context) pairs from parse trees. We “collapsed" nodes with prepositions and notated inverse relations separately, e.g., “dog barks" emits two tuples: (barks, nsubj_dog) and (dog, nsubj INLINEFORM0 _barks). We filter words and contexts that appear fewer than 100 times, resulting in INLINEFORM1 173k words and 1M contexts. We trained 300d vectors using word2vecf with default parameters. ### Setup
We compared our proposed approaches to a standard CNN that exploits a single set of word embeddings BIBREF3 . We also compared to a baseline of simply concatenating embeddings for each word to form long vector inputs. We refer to this as Concatenation-CNN C-CNN. For all multiple embedding approaches (C-CNN, MG-CNN and MGNC-CNN), we explored two combined sets of embedding: word2vec+Glove, and word2vec+syntactic, and one three sets of embedding: word2vec+Glove+syntactic. For all models, we tuned the l2 norm constraint INLINEFORM0 over the range INLINEFORM1 on a validation set. For instantiations of MGNC-CNN in which we exploited two embeddings, we tuned both INLINEFORM2 , and INLINEFORM3 ; where we used three embedding sets, we tuned INLINEFORM4 and INLINEFORM5 . We used standard train/test splits for those datasets that had them. Otherwise, we performed 10-fold cross validation, creating nested development sets with which to tune hyperparameters. For all experiments we used filters sizes of 3, 4 and 5 and we created 100 feature maps for each filter size. We applied 1 max-pooling and dropout (rate: 0.5) at the classification layer. For training we used back-propagation in mini-batches and used AdaDelta as the stochastic gradient descent (SGD) update rule, and set mini-batch size as 50. In this work, we treat word embeddings as part of the parameters of the model, and update them as well during training. In all our experiments, we only tuned the max norm constraint(s), fixing all other hyperparameters. ### Results and Discussion
We repeated each experiment 10 times and report the mean and ranges across these. This replication is important because training is stochastic and thus introduces variance in performance BIBREF4 . Results are shown in Table TABREF2 , and the corresponding best norm constraint value is shown in Table TABREF2 . We also show results on Subj, SST-1 and SST-2 achieved by the more complex model of BIBREF11 for comparison; this represents the state-of-the-art on the three datasets other than TREC. We can see that MGNC-CNN and MG-CNN always outperform baseline methods (including C-CNN), and MGNC-CNN is usually better than MG-CNN. And on the Subj dataset, MG-CNN actually achieves slightly better results than BIBREF11 , with far less complexity and required training time (MGNC-CNN performs comparably, although no better, here). On the TREC dataset, the best-ever accuracy we are aware of is 96.0% BIBREF21 , which falls within the range of the result of our MGNC-CNN model with three word embeddings. On the irony dataset, our model with three embeddings achieves 4% improvement (in terms of AUC) compared to the baseline model. On SST-1 and SST-2, our model performs slightly worse than BIBREF11 . However, we again note that their performance is achieved using a much more complex model which involves pre-training and mutual-learning steps. This model takes days to train, whereas our model requires on the order of an hour. We note that the method proposed by Astudillo et al. astudillo2015learning is able to accommodate multiple embedding sets with different dimensions by projecting the original word embeddings into a lower-dimensional space. However, this work requires training the optimal projection matrix on laebled data first, which again incurs large overhead. Of course, our model also has its own limitations: in MGNC-CNN, we need to tune the norm constraint hyperparameter for all the word embeddings. As the number of word embedding increases, this will increase the running time. However, this tuning procedure is embarrassingly parallel. ### Conclusions
We have proposed MGNC-CNN: a simple, flexible CNN architecture for sentence classification that can exploit multiple, variable sized word embeddings. We demonstrated that this consistently achieves better results than a baseline architecture that exploits only a single set of word embeddings, and also a naive concatenation approach to capitalizing on multiple embeddings. Furthermore, our results are comparable to those achieved with a recently proposed model BIBREF11 that is much more complex. However, our simple model is easy to implement and requires an order of magnitude less training time. Furthermore, our model is much more flexible than previous approaches, because it can accommodate variable-size word embeddings. ### Acknowledgments
This work was supported in part by the Army Research Office (grant W911NF-14-1-0442) and by The Foundation for Science and Technology, Portugal (grant UTAP-EXPL/EEIESS/0031/2014). This work was also made possible by the support of the Texas Advanced Computer Center (TACC) at UT Austin. Figure 1: Illustration of MG-CNN and MGNC-CNN. The filters applied to the respective embeddings are completely independent. MG-CNN applies a max norm constraint to o, while MGNC-CNN applies max norm constraints on o1 and o2 independently (group regularization). Note that one may easily extend the approach to handle more than two embeddings at once. Table 1: Results mean (min, max) achieved with each method. w2v:word2vec. Glv:GloVe. Syn: Syntactic embedding. Note that we experiment with using two and three sets of embeddings jointly, e.g., w2v+Syn+Glv indicates that we use all three of these. Table 2: Best λ2 value on the validation set for each method w2v:word2vec. Glv:GloVe. Syn: Syntactic embedding. | SST-1, SST-2, Subj , TREC , Irony |
What can you conclude about Retief's character?
A. He is gullible and easily tricked.
B. He is firm but can be harsh.
C. He has a soft spot for few in his life.
D. He can greedy and demanding.
| CULTURAL EXCHANGE BY KEITH LAUMER It was a simple student exchange—but Retief gave them more of an education than they expected! [Transcriber's Note: This etext was produced from Worlds of If Science Fiction, September 1962. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed.] I Second Secretary Magnan took his green-lined cape and orange-feathered beret from the clothes tree. "I'm off now, Retief," he said. "I hope you'll manage the administrative routine during my absence without any unfortunate incidents." "That seems a modest enough hope," Retief said. "I'll try to live up to it." "I don't appreciate frivolity with reference to this Division," Magnan said testily. "When I first came here, the Manpower Utilization Directorate, Division of Libraries and Education was a shambles. I fancy I've made MUDDLE what it is today. Frankly, I question the wisdom of placing you in charge of such a sensitive desk, even for two weeks. But remember. Yours is purely a rubber-stamp function." "In that case, let's leave it to Miss Furkle. I'll take a couple of weeks off myself. With her poundage, she could bring plenty of pressure to bear." "I assume you jest, Retief," Magnan said sadly. "I should expect even you to appreciate that Bogan participation in the Exchange Program may be the first step toward sublimation of their aggressions into more cultivated channels." "I see they're sending two thousand students to d'Land," Retief said, glancing at the Memo for Record. "That's a sizable sublimation." Magnan nodded. "The Bogans have launched no less than four military campaigns in the last two decades. They're known as the Hoodlums of the Nicodemean Cluster. Now, perhaps, we shall see them breaking that precedent and entering into the cultural life of the Galaxy." "Breaking and entering," Retief said. "You may have something there. But I'm wondering what they'll study on d'Land. That's an industrial world of the poor but honest variety." "Academic details are the affair of the students and their professors," Magnan said. "Our function is merely to bring them together. See that you don't antagonize the Bogan representative. This will be an excellent opportunity for you to practice your diplomatic restraint—not your strong point, I'm sure you'll agree." A buzzer sounded. Retief punched a button. "What is it, Miss Furkle?" "That—bucolic person from Lovenbroy is here again." On the small desk screen, Miss Furkle's meaty features were compressed in disapproval. "This fellow's a confounded pest. I'll leave him to you, Retief," Magnan said. "Tell him something. Get rid of him. And remember: here at Corps HQ, all eyes are upon you." "If I'd thought of that, I'd have worn my other suit," Retief said. Magnan snorted and passed from view. Retief punched Miss Furkle's button. "Send the bucolic person in." A tall broad man with bronze skin and gray hair, wearing tight trousers of heavy cloth, a loose shirt open at the neck and a short jacket, stepped into the room. He had a bundle under his arm. He paused at sight of Retief, looked him over momentarily, then advanced and held out his hand. Retief took it. For a moment the two big men stood, face to face. The newcomer's jaw muscles knotted. Then he winced. Retief dropped his hand and motioned to a chair. "That's nice knuckle work, mister," the stranger said, massaging his hand. "First time anybody ever did that to me. My fault though. I started it, I guess." He grinned and sat down. "What can I do for you?" Retief said. "You work for this Culture bunch, do you? Funny. I thought they were all ribbon-counter boys. Never mind. I'm Hank Arapoulous. I'm a farmer. What I wanted to see you about was—" He shifted in his chair. "Well, out on Lovenbroy we've got a serious problem. The wine crop is just about ready. We start picking in another two, three months. Now I don't know if you're familiar with the Bacchus vines we grow...?" "No," Retief said. "Have a cigar?" He pushed a box across the desk. Arapoulous took one. "Bacchus vines are an unusual crop," he said, puffing the cigar alight. "Only mature every twelve years. In between, the vines don't need a lot of attention, so our time's mostly our own. We like to farm, though. Spend a lot of time developing new forms. Apples the size of a melon—and sweet—" "Sounds very pleasant," Retief said. "Where does the Libraries and Education Division come in?" Arapoulous leaned forward. "We go in pretty heavy for the arts. Folks can't spend all their time hybridizing plants. We've turned all the land area we've got into parks and farms. Course, we left some sizable forest areas for hunting and such. Lovenbroy's a nice place, Mr. Retief." "It sounds like it, Mr. Arapoulous. Just what—" "Call me Hank. We've got long seasons back home. Five of 'em. Our year's about eighteen Terry months. Cold as hell in winter; eccentric orbit, you know. Blue-black sky, stars visible all day. We do mostly painting and sculpture in the winter. Then Spring; still plenty cold. Lots of skiing, bob-sledding, ice skating; and it's the season for woodworkers. Our furniture—" "I've seen some of your furniture," Retief said. "Beautiful work." Arapoulous nodded. "All local timbers too. Lots of metals in our soil and those sulphates give the woods some color, I'll tell you. Then comes the Monsoon. Rain—it comes down in sheets. But the sun's getting closer. Shines all the time. Ever seen it pouring rain in the sunshine? That's the music-writing season. Then summer. Summer's hot. We stay inside in the daytime and have beach parties all night. Lots of beach on Lovenbroy; we're mostly islands. That's the drama and symphony time. The theatres are set up on the sand, or anchored off-shore. You have the music and the surf and the bonfires and stars—we're close to the center of a globular cluster, you know...." "You say it's time now for the wine crop?" "That's right. Autumn's our harvest season. Most years we have just the ordinary crops. Fruit, grain, that kind of thing; getting it in doesn't take long. We spend most of the time on architecture, getting new places ready for the winter or remodeling the older ones. We spend a lot of time in our houses. We like to have them comfortable. But this year's different. This is Wine Year." Arapoulous puffed on his cigar, looked worriedly at Retief. "Our wine crop is our big money crop," he said. "We make enough to keep us going. But this year...." "The crop isn't panning out?" "Oh, the crop's fine. One of the best I can remember. Course, I'm only twenty-eight; I can't remember but two other harvests. The problem's not the crop." "Have you lost your markets? That sounds like a matter for the Commercial—" "Lost our markets? Mister, nobody that ever tasted our wines ever settled for anything else!" "It sounds like I've been missing something," said Retief. "I'll have to try them some time." Arapoulous put his bundle on the desk, pulled off the wrappings. "No time like the present," he said. Retief looked at the two squat bottles, one green, one amber, both dusty, with faded labels, and blackened corks secured by wire. "Drinking on duty is frowned on in the Corps, Mr. Arapoulous," he said. "This isn't drinking . It's just wine." Arapoulous pulled the wire retainer loose, thumbed the cork. It rose slowly, then popped in the air. Arapoulous caught it. Aromatic fumes wafted from the bottle. "Besides, my feelings would be hurt if you didn't join me." He winked. Retief took two thin-walled glasses from a table beside the desk. "Come to think of it, we also have to be careful about violating quaint native customs." Arapoulous filled the glasses. Retief picked one up, sniffed the deep rust-colored fluid, tasted it, then took a healthy swallow. He looked at Arapoulous thoughtfully. "Hmmm. It tastes like salted pecans, with an undercurrent of crusted port." "Don't try to describe it, Mr. Retief," Arapoulous said. He took a mouthful of wine, swished it around his teeth, swallowed. "It's Bacchus wine, that's all. Nothing like it in the Galaxy." He pushed the second bottle toward Retief. "The custom back home is to alternate red wine and black." Retief put aside his cigar, pulled the wires loose, nudged the cork, caught it as it popped up. "Bad luck if you miss the cork," Arapoulous said, nodding. "You probably never heard about the trouble we had on Lovenbroy a few years back?" "Can't say that I did, Hank." Retief poured the black wine into two fresh glasses. "Here's to the harvest." "We've got plenty of minerals on Lovenbroy," Arapoulous said, swallowing wine. "But we don't plan to wreck the landscape mining 'em. We like to farm. About ten years back some neighbors of ours landed a force. They figured they knew better what to do with our minerals than we did. Wanted to strip-mine, smelt ore. We convinced 'em otherwise. But it took a year, and we lost a lot of men." "That's too bad," Retief said. "I'd say this one tastes more like roast beef and popcorn over a Riesling base." "It put us in a bad spot," Arapoulous went on. "We had to borrow money from a world called Croanie. Mortgaged our crops. Had to start exporting art work too. Plenty of buyers, but it's not the same when you're doing it for strangers." "Say, this business of alternating drinks is the real McCoy," Retief said. "What's the problem? Croanie about to foreclose?" "Well, the loan's due. The wine crop would put us in the clear. But we need harvest hands. Picking Bacchus grapes isn't a job you can turn over to machinery—and anyway we wouldn't if we could. Vintage season is the high point of living on Lovenbroy. Everybody joins in. First, there's the picking in the fields. Miles and miles of vineyards covering the mountain sides, and crowding the river banks, with gardens here and there. Big vines, eight feet high, loaded with fruit, and deep grass growing between. The wine-carriers keep on the run, bringing wine to the pickers. There's prizes for the biggest day's output, bets on who can fill the most baskets in an hour.... The sun's high and bright, and it's just cool enough to give you plenty of energy. Come nightfall, the tables are set up in the garden plots, and the feast is laid on: roast turkeys, beef, hams, all kinds of fowl. Big salads. Plenty of fruit. Fresh-baked bread ... and wine, plenty of wine. The cooking's done by a different crew each night in each garden, and there's prizes for the best crews. "Then the wine-making. We still tramp out the vintage. That's mostly for the young folks but anybody's welcome. That's when things start to get loosened up. Matter of fact, pretty near half our young-uns are born after a vintage. All bets are off then. It keeps a fellow on his toes though. Ever tried to hold onto a gal wearing nothing but a layer of grape juice?" "Never did," Retief said. "You say most of the children are born after a vintage. That would make them only twelve years old by the time—" "Oh, that's Lovenbroy years; they'd be eighteen, Terry reckoning." "I was thinking you looked a little mature for twenty-eight," Retief said. "Forty-two, Terry years," Arapoulous said. "But this year it looks bad. We've got a bumper crop—and we're short-handed. If we don't get a big vintage, Croanie steps in. Lord knows what they'll do to the land. Then next vintage time, with them holding half our grape acreage—" "You hocked the vineyards?" "Yep. Pretty dumb, huh? But we figured twelve years was a long time." "On the whole," Retief said, "I think I prefer the black. But the red is hard to beat...." "What we figured was, maybe you Culture boys could help us out. A loan to see us through the vintage, enough to hire extra hands. Then we'd repay it in sculpture, painting, furniture—" "Sorry, Hank. All we do here is work out itineraries for traveling side-shows, that kind of thing. Now, if you needed a troop of Groaci nose-flute players—" "Can they pick grapes?" "Nope. Anyway, they can't stand the daylight. Have you talked this over with the Labor Office?" "Sure did. They said they'd fix us up with all the electronics specialists and computer programmers we wanted—but no field hands. Said it was what they classified as menial drudgery; you'd have thought I was trying to buy slaves." The buzzer sounded. Miss Furkle's features appeared on the desk screen. "You're due at the Intergroup Council in five minutes," she said. "Then afterwards, there are the Bogan students to meet." "Thanks." Retief finished his glass, stood. "I have to run, Hank," he said. "Let me think this over. Maybe I can come up with something. Check with me day after tomorrow. And you'd better leave the bottles here. Cultural exhibits, you know." II As the council meeting broke up, Retief caught the eye of a colleague across the table. "Mr. Whaffle, you mentioned a shipment going to a place called Croanie. What are they getting?" Whaffle blinked. "You're the fellow who's filling in for Magnan, over at MUDDLE," he said. "Properly speaking, equipment grants are the sole concern of the Motorized Equipment Depot, Division of Loans and Exchanges." He pursed his lips. "However, I suppose there's no harm in telling you. They'll be receiving heavy mining equipment." "Drill rigs, that sort of thing?" "Strip mining gear." Whaffle took a slip of paper from a breast pocket, blinked at it. "Bolo Model WV/1 tractors, to be specific. Why is MUDDLE interested in MEDDLE's activities?" "Forgive my curiosity, Mr. Whaffle. It's just that Croanie cropped up earlier today. It seems she holds a mortgage on some vineyards over on—" "That's not MEDDLE's affair, sir," Whaffle cut in. "I have sufficient problems as Chief of MEDDLE without probing into MUDDLE'S business." "Speaking of tractors," another man put in, "we over at the Special Committee for Rehabilitation and Overhaul of Under-developed Nations' General Economies have been trying for months to get a request for mining equipment for d'Land through MEDDLE—" "SCROUNGE was late on the scene," Whaffle said. "First come, first served. That's our policy at MEDDLE. Good day, gentlemen." He strode off, briefcase under his arm. "That's the trouble with peaceful worlds," the SCROUNGE committeeman said. "Boge is a troublemaker, so every agency in the Corps is out to pacify her. While my chance to make a record—that is, assist peace-loving d'Land—comes to naught." He shook his head. "What kind of university do they have on d'Land?" asked Retief. "We're sending them two thousand exchange students. It must be quite an institution." "University? D'Land has one under-endowed technical college." "Will all the exchange students be studying at the Technical College?" "Two thousand students? Hah! Two hundred students would overtax the facilities of the college." "I wonder if the Bogans know that?" "The Bogans? Why, most of d'Land's difficulties are due to the unwise trade agreement she entered into with Boge. Two thousand students indeed!" He snorted and walked away. Retief stopped by the office to pick up a short cape, then rode the elevator to the roof of the 230-story Corps HQ building and hailed a cab to the port. The Bogan students had arrived early. Retief saw them lined up on the ramp waiting to go through customs. It would be half an hour before they were cleared through. He turned into the bar and ordered a beer. A tall young fellow on the next stool raised his glass. "Happy days," he said. "And nights to match." "You said it." He gulped half his beer. "My name's Karsh. Mr. Karsh. Yep, Mr. Karsh. Boy, this is a drag, sitting around this place waiting...." "You meeting somebody?" "Yeah. Bunch of babies. Kids. How they expect—Never mind. Have one on me." "Thanks. You a Scoutmaster?" "I'll tell you what I am. I'm a cradle-robber. You know—" he turned to Retief—"not one of those kids is over eighteen." He hiccupped. "Students, you know. Never saw a student with a beard, did you?" "Lots of times. You're meeting the students, are you?" The young fellow blinked at Retief. "Oh, you know about it, huh?" "I represent MUDDLE." Karsh finished his beer, ordered another. "I came on ahead. Sort of an advance guard for the kids. I trained 'em myself. Treated it like a game, but they can handle a CSU. Don't know how they'll act under pressure. If I had my old platoon—" He looked at his beer glass, pushed it back. "Had enough," he said. "So long, friend. Or are you coming along?" Retief nodded. "Might as well." At the exit to the Customs enclosure, Retief watched as the first of the Bogan students came through, caught sight of Karsh and snapped to attention, his chest out. "Drop that, mister," Karsh snapped. "Is that any way for a student to act?" The youth, a round-faced lad with broad shoulders, grinned. "Heck, no," he said. "Say, uh, Mr. Karsh, are we gonna get to go to town? We fellas were thinking—" "You were, hah? You act like a bunch of school kids! I mean ... no! Now line up!" "We have quarters ready for the students," Retief said. "If you'd like to bring them around to the west side, I have a couple of copters laid on." "Thanks," said Karsh. "They'll stay here until take-off time. Can't have the little dears wandering around loose. Might get ideas about going over the hill." He hiccupped. "I mean they might play hookey." "We've scheduled your re-embarkation for noon tomorrow. That's a long wait. MUDDLE's arranged theater tickets and a dinner." "Sorry," Karsh said. "As soon as the baggage gets here, we're off." He hiccupped again. "Can't travel without our baggage, y'know." "Suit yourself," Retief said. "Where's the baggage now?" "Coming in aboard a Croanie lighter." "Maybe you'd like to arrange for a meal for the students here." "Sure," Karsh said. "That's a good idea. Why don't you join us?" Karsh winked. "And bring a few beers." "Not this time," Retief said. He watched the students, still emerging from Customs. "They seem to be all boys," he commented. "No female students?" "Maybe later," Karsh said. "You know, after we see how the first bunch is received." Back at the MUDDLE office, Retief buzzed Miss Furkle. "Do you know the name of the institution these Bogan students are bound for?" "Why, the University at d'Land, of course." "Would that be the Technical College?" Miss Furkle's mouth puckered. "I'm sure I've never pried into these details." "Where does doing your job stop and prying begin, Miss Furkle?" Retief said. "Personally, I'm curious as to just what it is these students are travelling so far to study—at Corps expense." "Mr. Magnan never—" "For the present. Miss Furkle, Mr. Magnan is vacationing. That leaves me with the question of two thousand young male students headed for a world with no classrooms for them ... a world in need of tractors. But the tractors are on their way to Croanie, a world under obligation to Boge. And Croanie holds a mortgage on the best grape acreage on Lovenbroy." "Well!" Miss Furkle snapped, small eyes glaring under unplucked brows. "I hope you're not questioning Mr. Magnan's wisdom!" "About Mr. Magnan's wisdom there can be no question," Retief said. "But never mind. I'd like you to look up an item for me. How many tractors will Croanie be getting under the MEDDLE program?" "Why, that's entirely MEDDLE business," Miss Furkle said. "Mr. Magnan always—" "I'm sure he did. Let me know about the tractors as soon as you can." Miss Furkle sniffed and disappeared from the screen. Retief left the office, descended forty-one stories, followed a corridor to the Corps Library. In the stacks he thumbed through catalogues, pored over indices. "Can I help you?" someone chirped. A tiny librarian stood at his elbow. "Thank you, ma'am," Retief said. "I'm looking for information on a mining rig. A Bolo model WV tractor." "You won't find it in the industrial section," the librarian said. "Come along." Retief followed her along the stacks to a well-lit section lettered ARMAMENTS. She took a tape from the shelf, plugged it into the viewer, flipped through and stopped at a squat armored vehicle. "That's the model WV," she said. "It's what is known as a continental siege unit. It carries four men, with a half-megaton/second firepower." "There must be an error somewhere," Retief said. "The Bolo model I want is a tractor. Model WV M-1—" "Oh, the modification was the addition of a bulldozer blade for demolition work. That must be what confused you." "Probably—among other things. Thank you." Miss Furkle was waiting at the office. "I have the information you wanted," she said. "I've had it for over ten minutes. I was under the impression you needed it urgently, and I went to great lengths—" "Sure," Retief said. "Shoot. How many tractors?" "Five hundred." "Are you sure?" Miss Furkle's chins quivered. "Well! If you feel I'm incompetent—" "Just questioning the possibility of a mistake, Miss Furkle. Five hundred tractors is a lot of equipment." "Was there anything further?" Miss Furkle inquired frigidly. "I sincerely hope not," Retief said. III Leaning back in Magnan's padded chair with power swivel and hip-u-matic concontour, Retief leafed through a folder labelled "CERP 7-602-Ba; CROANIE (general)." He paused at a page headed Industry. Still reading, he opened the desk drawer, took out the two bottles of Bacchus wine and two glasses. He poured an inch of wine into each and sipped the black wine meditatively. It would be a pity, he reflected, if anything should interfere with the production of such vintages.... Half an hour later he laid the folder aside, keyed the phone and put through a call to the Croanie Legation. He asked for the Commercial Attache. "Retief here, Corps HQ," he said airily. "About the MEDDLE shipment, the tractors. I'm wondering if there's been a slip up. My records show we're shipping five hundred units...." "That's correct. Five hundred." Retief waited. "Ah ... are you there, Retief?" "I'm still here. And I'm still wondering about the five hundred tractors." "It's perfectly in order. I thought it was all settled. Mr. Whaffle—" "One unit would require a good-sized plant to handle its output," Retief said. "Now Croanie subsists on her fisheries. She has perhaps half a dozen pint-sized processing plants. Maybe, in a bind, they could handle the ore ten WV's could scrape up ... if Croanie had any ore. It doesn't. By the way, isn't a WV a poor choice as a mining outfit? I should think—" "See here, Retief! Why all this interest in a few surplus tractors? And in any event, what business is it of yours how we plan to use the equipment? That's an internal affair of my government. Mr. Whaffle—" "I'm not Mr. Whaffle. What are you going to do with the other four hundred and ninety tractors?" "I understood the grant was to be with no strings attached!" "I know it's bad manners to ask questions. It's an old diplomatic tradition that any time you can get anybody to accept anything as a gift, you've scored points in the game. But if Croanie has some scheme cooking—" "Nothing like that, Retief. It's a mere business transaction." "What kind of business do you do with a Bolo WV? With or without a blade attached, it's what's known as a continental siege unit." "Great Heavens, Retief! Don't jump to conclusions! Would you have us branded as warmongers? Frankly—is this a closed line?" "Certainly. You may speak freely." "The tractors are for transshipment. We've gotten ourselves into a difficult situation, balance-of-payments-wise. This is an accommodation to a group with which we have rather strong business ties." "I understand you hold a mortgage on the best land on Lovenbroy," Retief said. "Any connection?" "Why ... ah ... no. Of course not, ha ha." "Who gets the tractors eventually?" "Retief, this is unwarranted interference!" "Who gets them?" "They happen to be going to Lovenbroy. But I scarcely see—" "And who's the friend you're helping out with an unauthorized transshipment of grant material?" "Why ... ah ... I've been working with a Mr. Gulver, a Bogan representative." "And when will they be shipped?" "Why, they went out a week ago. They'll be half way there by now. But look here, Retief, this isn't what you're thinking!" "How do you know what I'm thinking? I don't know myself." Retief rang off, buzzed the secretary. "Miss Furkle, I'd like to be notified immediately of any new applications that might come in from the Bogan Consulate for placement of students." "Well, it happens, by coincidence, that I have an application here now. Mr. Gulver of the Consulate brought it in." "Is Mr. Gulver in the office? I'd like to see him." "I'll ask him if he has time." "Great. Thanks." It was half a minute before a thick-necked red-faced man in a tight hat walked in. He wore an old-fashioned suit, a drab shirt, shiny shoes with round toes and an ill-tempered expression. "What is it you wish?" he barked. "I understood in my discussions with the other ... ah ... civilian there'd be no further need for these irritating conferences." "I've just learned you're placing more students abroad, Mr. Gulver. How many this time?" "Two thousand." "And where will they be going?" "Croanie. It's all in the application form I've handed in. Your job is to provide transportation." "Will there be any other students embarking this season?" "Why ... perhaps. That's Boge's business." Gulver looked at Retief with pursed lips. "As a matter of fact, we had in mind dispatching another two thousand to Featherweight." "Another under-populated world—and in the same cluster, I believe," Retief said. "Your people must be unusually interested in that region of space." "If that's all you wanted to know, I'll be on my way. I have matters of importance to see to." After Gulver left, Retief called Miss Furkle in. "I'd like to have a break-out of all the student movements that have been planned under the present program," he said. "And see if you can get a summary of what MEDDLE has been shipping lately." Miss Furkle compressed her lips. "If Mr. Magnan were here, I'm sure he wouldn't dream of interfering in the work of other departments. I ... overheard your conversation with the gentleman from the Croanie Legation—" "The lists, Miss Furkle." "I'm not accustomed," Miss Furkle said, "to intruding in matters outside our interest cluster." "That's worse than listening in on phone conversations, eh? But never mind. I need the information, Miss Furkle." "Loyalty to my Chief—" "Loyalty to your pay-check should send you scuttling for the material I've asked for," Retief said. "I'm taking full responsibility. Now scat." The buzzer sounded. Retief flipped a key. "MUDDLE, Retief speaking...." Arapoulous's brown face appeared on the desk screen. "How-do, Retief. Okay if I come up?" "Sure, Hank. I want to talk to you." In the office, Arapoulous took a chair. "Sorry if I'm rushing you, Retief," he said. "But have you got anything for me?" Retief waved at the wine bottles. "What do you know about Croanie?" "Croanie? Not much of a place. Mostly ocean. All right if you like fish, I guess. We import our seafood from there. Nice prawns in monsoon time. Over a foot long." "You on good terms with them?" "Sure, I guess so. Course, they're pretty thick with Boge." "So?" "Didn't I tell you? Boge was the bunch that tried to take us over here a dozen years back. They'd've made it too, if they hadn't had a lot of bad luck. Their armor went in the drink, and without armor they're easy game." Miss Furkle buzzed. "I have your lists," she said shortly. "Bring them in, please." The secretary placed the papers on the desk. Arapoulous caught her eye and grinned. She sniffed and marched from the room. "What that gal needs is a slippery time in the grape mash," Arapoulous observed. Retief thumbed through the papers, pausing to read from time to time. He finished and looked at Arapoulous. "How many men do you need for the harvest, Hank?" Retief inquired. Arapoulous sniffed his wine glass and looked thoughtful. "A hundred would help," he said. "A thousand would be better. Cheers." "What would you say to two thousand?" "Two thousand? Retief, you're not fooling?" "I hope not." He picked up the phone, called the Port Authority, asked for the dispatch clerk. "Hello, Jim. Say, I have a favor to ask of you. You know that contingent of Bogan students. They're traveling aboard the two CDT transports. I'm interested in the baggage that goes with the students. Has it arrived yet? Okay, I'll wait." Jim came back to the phone. "Yeah, Retief, it's here. Just arrived. But there's a funny thing. It's not consigned to d'Land. It's ticketed clear through to Lovenbroy." "Listen, Jim," Retief said. "I want you to go over to the warehouse and take a look at that baggage for me." Retief waited while the dispatch clerk carried out the errand. The level in the two bottles had gone down an inch when Jim returned to the phone. "Hey, I took a look at that baggage, Retief. Something funny going on. Guns. 2mm needlers, Mark XII hand blasters, power pistols—" "It's okay, Jim. Nothing to worry about. Just a mix-up. Now, Jim, I'm going to ask you to do something more for me. I'm covering for a friend. It seems he slipped up. I wouldn't want word to get out, you understand. I'll send along a written change order in the morning that will cover you officially. Meanwhile, here's what I want you to do...." Retief gave instructions, then rang off and turned to Arapoulous. "As soon as I get off a couple of TWX's, I think we'd better get down to the port, Hank. I think I'd like to see the students off personally." | B. He is firm but can be harsh. |
Which of the following most likely happened to Krugman after these letters?
A. Krugman wrote an official apology to the writers.
B. Krugman wrote another book about increasing returns.
C. Krugman quit writing in newspapers.
D. Krugman lost credibility among his colleagues.
| Krugman's Life of Brian Where it all started: Paul Krugman's "The Legend of Arthur." Letter from John Cassidy Paul Krugman replies to John Cassidy Letter from M. Mitchell Waldrop Paul Krugman replies to M. Mitchell Waldrop Letter from Kenneth J. Arrow Letter from Ted C. Fishman David Warsh's July 3, 1994, Boston Globe Letter from John Cassidy: Paul Krugman loves to berate journalists for their ignorance of economics, particularly his economics, but on this occasion, I fear, his logic is more addled than usual. I am reluctant to dignify his hatchet job with a lengthy reply, but some of his claims are so defamatory that they should be addressed, if only for the record. 1) Krugman claims that my opening sentence--"In a way, Bill Gates's current troubles with the Justice Department grew out of an economics seminar that took place thirteen years ago, at Harvard's John F. Kennedy School of Government"--is "pure fiction." Perhaps so, but in that case somebody should tell this to Joel Klein, the assistant attorney general in charge of the antitrust division. When I interviewed Klein for my piece about the Microsoft case, he singled out Brian Arthur as the economist who has most influenced his thinking about the way in which high-technology markets operate. It was Klein's words, not those of Arthur, that prompted me to use Arthur in the lead of the story. 2) Krugman wrote: "Cassidy's article tells the story of how Stanford Professor Brian Arthur came up with the idea of increasing returns." I wrote no such thing, and Arthur has never, to my knowledge, claimed any such thing. The notion of increasing returns has been around since Adam Smith, and it was written about at length by Alfred Marshall in 1890. What I did say in my article was that increasing returns was largely ignored by mainstream economists for much of the postwar era, a claim that simply isn't controversial. (As Krugman notes, one reason for this was technical, not ideological. Allowing for the possibility of increasing returns tends to rob economic models of two properties that economists cherish: simplicity and determinism. As long ago as 1939, Sir John Hicks, one of the founders of modern economics, noted that increasing returns, if tolerated, could lead to the "wreckage" of a large part of economic theory.) 3) Pace Krugman, I also did not claim that Arthur bears principal responsibility for the rediscovery of increasing returns by economists in the 1970s and 1980s. As Krugman notes, several scholars (himself included) who were working in the fields of game theory and international trade published articles incorporating increasing returns before Arthur did. My claim was simply that Arthur applied increasing returns to high-technology markets, and that his work influenced how other economists and government officials think about these markets. Krugman apart, virtually every economist I have spoken to, including Daniel Rubinfeld, a former Berkeley professor who is now the chief economist at the Justice Department's antitrust division, told me this was the case. (Rubinfeld also mentioned several other economists who did influential work, and I cited three of them in the article.) 4) Krugman appears to suggest that I made up some quotes, a charge that, if it came from a more objective source, I would consider to be a serious matter. In effect, he is accusing Brian Arthur, a man he calls a "nice guy," of being a fabricator or a liar. The quotes in question came from Arthur, and they were based on his recollections of two meetings that he attended some years ago. After Krugman's article appeared, the Santa Fe professor called me to say that he still recalled the meetings in question as I described them. Krugman, as he admits, wasn't present at either of the meetings. 5) For a man who takes his own cogitations extremely seriously, Krugman is remarkably cavalier about attributing motives and beliefs to others. "Cassidy has made it clear in earlier writing that he does not like mainstream economists, and he may have been overly eager to accept a story that puts them in a bad light," he pronounces. I presume this statement refers to a critical piece I wrote in 1996 about the direction that economic research, principally macroeconomic research, has taken over the past two decades. In response to that article, I received dozens of messages of appreciation from mainstream economists, including from two former presidents of the American Economic Association. Among the sources quoted in that piece were the then-chairman of the White House Council of Economic Advisers (Joseph Stiglitz), a governor of the Federal Reserve Board (Laurence Meyer), and a well-known Harvard professor (Gregory Mankiw). To claim, as Krugman does, that I "don't like mainstream economists" and that I am out to denigrate their work is malicious hogwash. The fact of the matter is that I spend much of my life reading the work of mainstream economists, speaking to them, and trying to find something they have written that might interest the general public. In my experience, most economists appreciate the attention. 6) I might attach more weight to Krugman's criticisms if I hadn't recently reread his informative 1994 book Peddling Prosperity , in which he devotes a chapter to the rediscovery of increasing returns by contemporary economists. Who are the first scholars Krugman mentions in his account? Paul David, an economic historian who wrote a famous paper about how the QWERTYUIOP typewriter keyboard evolved and, you guessed it, Brian Arthur. "Why QWERTYUIOP?" Krugman wrote. "In the early 1980s, Paul David and his Stanford colleague Brian Arthur asked that question, and quickly realized that it led them into surprisingly deep waters. ... What Paul David, Brian Arthur, and a growing number of other economists began to realize in the late seventies and early eighties was that stories like that of the typewriter keyboard are, in fact, pervasive in the economy." Evidently, Krugman felt four years ago that Arthur's contribution was important enough to merit a prominent mention in his book. Now, he dismisses the same work, saying it "didn't tell me anything that I didn't already know." Doubtless, this change in attitude on Krugman's part is unconnected to the fact that Arthur has started to receive some public recognition. The eminent MIT professor, whose early academic work received widespread media attention, is far too generous a scholar to succumb to such pettiness. --John Cassidy Paul Krugman replies to John Cassidy: I think that David Warsh's 1994 in the Boston Globe says it all. If other journalists would do as much homework as he did, I wouldn't have had to write that article. Letter from M. Mitchell Waldrop: Thanks to Paul Krugman for his lament about credulous reporters who refuse to let facts stand in the way of a good story ("The Legend of Arthur"). As a professional journalist, I found his points well taken--even when he cites my own book, Complexity as a classic example of the gullibility genre. Among many other things, Complexity tells the story of the Irish-born economist Brian Arthur and how he came to champion a principle known as "increasing returns." The recent New Yorker article explains how that principle has since become the intellectual foundation of the Clinton administration's antitrust case against Microsoft. Krugman's complaint is that the popular press--including Complexity and The New Yorker --is now hailing Brian Arthur as the originator of increasing returns, even though Krugman and many others had worked on the idea long before Arthur did. I leave it for others to decide whether I was too gullible in writing Complexity . For the record, however, I would like to inject a few facts into Krugman's story, which he summarizes nicely in the final paragraph: When Waldrop's book came out, I wrote him as politely as I could, asking exactly how he had managed to come up with his version of events. He did, to his credit, write back. He explained that while he had become aware of some other people working on increasing returns, trying to put them in would have pulled his story line out of shape. ... So what we really learn from the legend of Arthur is that some journalists like a good story too much to find out whether it is really true. Now, I will admit to many sins, not the least of them being a profound ignorance of graduate-level economics; I spent my graduate-school career in the physics department instead, writing a Ph.D. dissertation on the quantum-field theory of elementary particle collisions at relativistic energies. However, I am not so ignorant of the canons of journalism (and of common sense) that I would take a plausible fellow like Brian Arthur at face value without checking up on him. During my research for Complexity I spoke to a number of economists about his work, including Nobel laureate Kenneth Arrow, co-creator of the General Equilibrium Theory of economics that Brian so eloquently criticizes. They generally agreed that Brian was a maverick in the field--and perhaps a bit too much in love with his own self-image as a misunderstood outsider--but basically sound. None of them warned me that he was usurping credit where credit was not due. Which brings me to Professor Krugman's letter, and my reply. I remember the exchange very well. Obviously, however, my reply failed to make clear what I was really trying to say. So I'll try again: a) During our interviews, Brian went out of his way to impress upon me that many other economists had done work in increasing returns--Paul Krugman among them. He was anxious that they be given due credit in anything I wrote. So was I. b) Accordingly, I included a passage in Complexity in which Brian does indeed describe what others had done in the field--Paul Krugman among them. Elsewhere in that same chapter, I tried to make it clear that the concept of increasing returns was already well known to Brian's professors at Berkeley, where he first learned of it. Indeed, I quote Brian pointing out that increasing returns had been extensively discussed by the great English economist Alfred Marshall in 1891. c) So, when I received Krugman's letter shortly after Complexity came out, I was puzzled: He was complaining that I hadn't referenced others in the increasing-returns field--Paul Krugman among them--although I had explicitly done so. d) But, when I checked the published text, I was chagrined to discover that the critical passage mentioning Krugman wasn't there. e) Only then did I realize what had happened. After I had submitted the manuscript, my editor at Simon & Schuster had suggested a number of cuts to streamline what was already a long and involved chapter on Brian's ideas. I accepted some of the cuts, and restored others--including (I thought) the passage that mentioned Krugman. In the rush to get Complexity to press, however, that passage somehow wound up on the cutting-room floor anyway, and I didn't notice until too late. That oversight was my fault entirely, not my editor's, and certainly not Brian Arthur's. I take full responsibility, I regret it, and--if Simon & Schuster only published an errata column--I would happily correct it publicly. However, contrary to what Professor Krugman implies, it was an oversight, not a breezy disregard of facts for the sake of a good story. --M. Mitchell Waldrop Washington Paul Krugman replies to M. Mitchell Waldrop: I am truly sorry that The New Yorker has not yet established a Web presence so that we could include a link directly to the Cassidy piece. However, you can get a pretty good idea of what the piece said by reading the summary of it presented in "Tasty Bits from the Technology Front." Cassidy did not present a story about one guy among many who worked on increasing returns. On the contrary: He presented a morality play in which a lonely hero struggled to make his ideas heard against the unified opposition of a narrow-minded profession both intellectually and politically conservative. As TBTF's host--not exactly a naive reader--put it, "These ideas were anathema to mainstream economists in 1984 when Arthur first tried to publish them." That morality play--not the question of who deserves credit--was the main point of my column, because it is a pure (and malicious) fantasy that has nonetheless become part of the story line people tell about increasing returns and its relationship to mainstream economics. The fact, which is easily documented, is that during the years that, according to the legend, increasing returns was unacceptable in mainstream economics, papers about increasing returns were in fact being cheerfully published by all the major journals. And as I pointed out in the chronology I provided with the article, even standard reference volumes like the Handbook of International Economics (published in 1984, the year Arthur supposedly met a blank wall of resistance) have long contained chapters on increasing returns. Whatever the reason that Arthur had trouble getting his own paper published, ideological rigidity had nothing to do with it. How did this fantasy come to be so widely believed? I am glad to hear that you tried to tell a more balanced story, Mr. Waldrop, even if sloppy paperwork kept it from seeing the light of day. And I am glad that you talked to Ken Arrow. But Nobel laureates, who have wide responsibilities and much on their mind, are not necessarily on top of what has been going on in research outside their usual field. I happen to know of one laureate who, circa 1991, was quite unaware that anyone had thought about increasing returns in either growth or trade. Did you try talking to anyone else--say, to one of the economists who are the straight men in the stories you tell? For example, your book starts with the story of Arthur's meeting in 1987 with Al Fishlow at Berkeley, in which Fishlow supposedly said, "We know that increasing returns can't exist"--and Arthur went away in despair over the unwillingness of economists to think the unthinkable. Did you call Fishlow to ask whether he said it, and what he meant? Since by 1987 Paul Romer's 1986 papers on increasing returns and growth had started an avalanche of derivative work, he was certainly joking--what he probably meant was "Oh no, not you too." And let me say that I simply cannot believe that you could have talked about increasing returns with any significant number of economists outside Santa Fe without Romer's name popping up in the first 30 seconds of every conversation--unless you were very selective about whom you talked to. And oh, by the way, there are such things as libraries, where you can browse actual economics journals and see what they contain. The point is that it's not just a matter of failing to cite a few more people. Your book, like the Cassidy article, didn't just tell the story of Brian Arthur; it also painted a picture of the economics profession, its intellectual bigotry and prejudice, which happens to be a complete fabrication (with some real, named people cast as villains) that somehow someone managed to sell you. I wonder who? Even more to the point: How did Cassidy come by his story? Is it possible that he completely misunderstood what Brian Arthur was saying--that the whole business about the seminar at Harvard where nobody would accept increasing returns, about the lonely struggle of Arthur in the face of ideological rigidity, even the quotation from Arthur about economists being unwilling to consider the possibility of imperfect markets because of the Cold War (give me a break!) were all in Cassidy's imagination? Let me say that I am actually quite grateful to Cassidy and The New Yorker . A number of people have long been furious about your book--for example, Victor Norman, whom you portrayed as the first of many economists too dumb or perhaps narrow-minded to understand Arthur's brilliant innovation. Norman e-mailed me to say that "I have read the tales from the Vienna woods before and had hoped that it could be cleared up by someone at some point." Yet up to now there was nothing anyone could do about the situation. The trouble was that while "heroic rebel defies orthodoxy" is a story so good that nobody even tries to check it out, "guy makes minor contribution to well-established field, proclaims himself its founder" is so boring as to be unpublishable. (David Warsh's 1994 series of columns in the Boston Globe on the increasing-returns revolution in economics, the basis for a forthcoming book from Harvard University Press, is far and away the best reporting on the subject, did include a sympathetic but devastating exposé of Arthur's pretensions--but to little effect. [Click to read Warsh on Arthur.]) Only now did I have a publishable story: "guy makes minor contribution to well-established field, portrays himself as heroic rebel--and The New Yorker believes him." Thank you, Mr. Cassidy. Letter from Kenneth J. Arrow: Paul Krugman's attack on Brian Arthur ("The Legend of Arthur") requires a correction of its misrepresentations of fact. Arthur is a reputable and significant scholar whose work is indeed having influence in the field of industrial organization and in particular public policy toward antitrust policy in high-tech industries. Krugman admits that he wrote the article because he was "just pissed off," not a very good state for a judicious statement of facts, as his column shows. His theme is stated in his first paragraph: "Cassidy's article [in The New Yorker of Jan. 12] tells the story of how Stanford Professor Brian Arthur came up with the idea of increasing returns." Cassidy, however, said nothing of the sort. The concept of increasing returns is indeed very old, and Cassidy at no point attributed that idea to Arthur. Indeed, the phrase "increasing returns" appears just once in Cassidy's article and then merely to say that Arthur had used the term while others refer to network externalities. Further, Arthur has never made any such preposterous claim at any other time. On the contrary, his papers have fully cited the history of the field and made references to the previous papers, including those of Paul Krugman. (See Arthur's papers collected in the volume Increasing Returns and Path Dependence in the Economy, especially his preface and my foreword for longer comments on Arthur's work in historic perspective. Click to see the foreword.) Hence, Krugman's whole attack is directed at a statement made neither by Arthur nor by Cassidy. Krugman has not read Cassidy's piece with any care nor has he bothered to review what Arthur has in fact said. What Cassidy in fact did in his article was to trace a line of influence between one of Arthur's early articles and the current claims of the Department of Justice against Microsoft. It appears that Cassidy based his article on several interviews, not just one. The point that Arthur has emphasized and which is influential in the current debates about antitrust policy is the dynamic implication of increasing returns. It is the concept of path-dependence, that small events, whether random or the result of corporate strategic choice, may have large consequences because of increasing returns of various kinds. Initial small advantages become magnified, for example, by creating a large installed base, and direct the future, possibly in an inefficient direction. Techniques of production may be locked in at an early stage. Similar considerations apply to regional development and learning. --Kenneth J. Arrow Nobel laureate and Joan Kenney professor of economics emeritus Stanford University Letter from Ted C. Fishman: After reading Paul Krugman vent his spleen against fellow economist Brian Arthur in "The Legend of Arthur," I couldn't help wondering whose reputation he was out to trash, Arthur's or his own. Krugman seems to fear a plot to deny economists their intellectual due. If one exists, Arthur is not a likely suspect. In a series of long interviews with me a year ago (for Worth magazine), I tried, vainly, to get Arthur to tell me how his ideas about increasing returns have encouraged a new strain of economic investigations. Despite much prodding, Arthur obliged only by placing himself in a long line of theorists dating back to Adam Smith and Alfred Marshall. I also found him disarmingly generous in giving credit to the biologists, physicists, and fellow economists who have helped advance his own thinking. Savvy to the journalist's quest for heroes, Arthur urged me to focus on his ideas, not his rank among his peers. Krugman has made a career out of telling other economists to pay better attention to the facts, yet as a chronicler of Arthur's career and inner life, Krugman seems to have listened only to his own demons. --Ted C. Fishman (For additional background on the history of "increasing returns" and Brian Arthur's standing in the field, click for David Warsh's July 3, 1994, Boston Globe article on Brian Arthur) | D. Krugman lost credibility among his colleagues. |
Which dataset do they use? | ### Introduction
An important property of human communication is that listeners can infer information beyond the literal meaning of an utterance. One well-studied type of inference is scalar inference BIBREF0, BIBREF1, whereby a listener who hears an utterance with a scalar item like some infers the negation of a stronger alternative with all: Chris ate some of the cookies. $\rightsquigarrow $ Chris ate some, but not all, of the cookies. Early accounts of scalar inferences (e.g., BIBREF2, BIBREF1, BIBREF3) considered them to arise by default unless specifically cancelled by the context. However, in a recent corpus study, degen2015investigating showed that there is much more variability in the strength of scalar inferences from some to not all than previously assumed. degen2015investigating further showed that this variability is not random and that several lexical, syntactic, and semantic/pragmatic features of context explain much of the variance in inference strength. Recent Bayesian game-theoretic models of pragmatic reasoning BIBREF4, BIBREF5, which are capable of integrating multiple linguistic cues with world knowledge, are able to correctly predict listeners' pragmatic inferences in many cases (e.g., BIBREF6, BIBREF7). These experimental and modeling results suggest that listeners integrate multiple linguistic and contextual cues in utterance interpretation, raising the question how listeners are able to draw these pragmatic inferences so quickly in interaction. This is an especially pressing problem considering that inference in Bayesian models of pragmatics is intractable when scaling up beyond toy domains to make predictions about arbitrary utterances. One possibility is that language users learn to use shortcuts to the inference (or lack thereof) by learning associations between the speaker's intention and surface-level cues present in the linguistic signal across many instances of encountering a scalar expression like some. In this work, we investigate whether it is possible to learn such associations between cues in the linguistic signal and speaker intentions by training neural network models to predict empirically elicited inference strength ratings from the linguistic input. In this enterprise we follow the recent successes of neural network models in predicting a range of linguistic phenomena such as long distance syntactic dependencies (e.g., BIBREF11, BIBREF12, BIBREF13, BIBREF14, BIBREF15), semantic entailments (e.g., BIBREF16, BIBREF17), acceptability judgements BIBREF18, factuality BIBREF19, and, to some extent, speaker commitment BIBREF20. In particular, we ask: How well can a neural network sentence encoder learn to predict human inference strength judgments for utterances with some? To what extent does such a model capture the qualitative effects of hand-mined contextual features previously identified as influencing inference strength? To address these questions, we first compare the performance of neural models that differ in the underlying word embedding model (GloVe, ELMo, or BERT) and in the sentence embedding model (LSTM, LSTM$+$attention). We then probe the best model's behavior through a regression analysis, an analysis of attention weights, and an analysis of predictions on manually constructed minimal sentence pairs. ### The dataset
We used the annotated dataset reported by degen2015investigating, a dataset of the utterances from the Switchboard corpus of telephone dialogues BIBREF21 that contain the word some. The dataset consists of 1,362 unique utterances with a noun phrase containing some (some-NP). For each example with a some-NP, degen2015investigating collected inference strength ratings from at least 10 participants recruited on Amazon's Mechanical Turk. Participants saw both the target utterance and ten utterances from the preceding discourse context. They then rated the similarity between the original utterance like (UNKREF8) and an utterance in which some was replaced with some, but not all like (UNKREF9), on a 7-point Likert scale with endpoints labeled “very different meaning” (1) and “same meaning” (7). Low similarity ratings thus indicate low inference strength, and high similarity ratings indicate high inference strength. I like, I like to read some of the philosophy stuff. I like, I like to read some, but not all, of the philosophy stuff. Using this corpus, degen2015investigating found that several linguistic and contextual factors influenced inference strength ratings, including the partitive form of, subjecthood, the previous mention of the embedded NP referent, determiner strength, and modification of the head noun. Partitive: (UNKREF10a-b) are example utterances from the corpus with and without partitive some-NPs, respectively. Values in parentheses indicate the mean inference strength rating for that item. On average, utterances with partitives yielded stronger inference ratings than ones without. I've seen some of them on repeat. (5.8) You sound like you have some small ones in the background. (1.5) Subjecthood: Utterances in which the some-NP appears in subject position, as in (UNKREF13a), yielded stronger inference ratings than utterances in which the some-NP appears in a different grammatical position, e.g., as a direct object as in (UNKREF13b). Some kids are really having it. (5.9) That would take some planning. (1.4) Previous mention: Discourse properties also have an effect on inference strength. A some-NP with a previously mentioned embedded NP referent yields stronger inferences than a some-NP whose embedded NP referent has not been previously mentioned. For example, (UNKREF16a) contains a some-NP in which them refers to previously mentioned Mission Impossible tape recordings, whereas planning in the some-NP in (UNKREF16b) has not been previously mentioned. I've seen some of them on repeats. (5.8) That would take some planning. (1.4) Modification: BIBREF22 also found a small effect of whether or not the head noun of the some-NP was modified such that some-NP with unmodified head nouns yielded slightly stronger inferences than those with modified head nouns. Determiner strength: Finally, the strength of the determiner some has traditionally been analyzed as having a weak, indefinite, non-presuppositional reading as well as a strong, quantificational, presuppositional reading BIBREF23, BIBREF24. While the weak/strong distinction has been notoriously hard to pin down BIBREF25, degen2015investigating used strength norms elicited independently for each item, which exploited the presuppositional nature of strong some: removing some (of) from utterances with weak some leads to higher ratings on a 7-point Likert scale from `different meaning' to `same meaning' than removing it from utterances with strong some. Items with stronger some – e.g., (UNKREF19a), strength 3.3 – yielded stronger inference ratings than items with weaker some – e.g., (UNKREF19b), strength 6.7. And some people don't vote. (5.2) Well, we could use some rain up here. (2.1) The quantitative findings from degen2015investigating are summarized in Figure FIGREF27, which shows in blue the regression coefficients for all predictors she considered (see the original paper for more detailed descriptions). For our experiments, we randomly split the dataset into a 70% training and 30% test set, resulting in 954 training items and 408 test items. ### Model
The objective of the model is to predict mean inference strength rating $i$ given an utterance (a sequence of words) $U = \lbrace w_1,w_2,...,w_N\rbrace $. While the original participant ratings were on a Likert scale from 1 to 7, we rescale these values to the interval $[0,1]$. Figure FIGREF22 shows the overall model architecture. The model is a sentence classification model akin to the model proposed by BIBREF26. The model first embeds the utterance tokens using pre-trained embedding models, and then forms a sentence representation by passing the embedded tokens through a 2-layer bidirectional LSTM network (biLSTM) BIBREF27 with dropout BIBREF28 followed by a self-attention mechanisms that provides a weighted average of the hidden states of the top-most biLSTM hidden states. This sentence representation is then passed through a transformation layer with a sigmoid activation function, which outputs the predicted score in the interval $[0,1]$. We rescale this predicted value to fall in the original interval $[1,7]$. We evaluated three word embedding models: GloVe, a static pre-trained word embedding matrix BIBREF29, and pre-trained contextual word embedding models in the form of English ELMo BIBREF30, BIBREF31 and English BERT BIBREF32, BIBREF33 models. We used the 100d GloVe embeddings, and we evaluated the 768d uncased BERT-base and 1024d BERT-large models. ### Experiments ::: Training
We used 5-fold cross-validation on the training data to optimize the following hyperparameters. Word embedding model: GloVe, ELMo, BERT-base, BERT-large. Output layer of word embedding models: $[1,3]$ for ELMo, $[1,12]$ for BERT-base, and $[1,24]$ for BERT-large. Number of training epochs: $[1,800]$. Dimension of LSTM hidden states: $\lbrace 100,200,400,800\rbrace $. Dropout rate in LSTM: $\lbrace 0.1,0.2,0.3,0.4\rbrace $. We first optimized the output layer of the word embedding model for each embedding model while keeping all other parameters fixed. We then optimized the other parameters for each embedding model by computing the average correlation between the model predictions and the human ratings across the five cross-validation folds. Architectural variants. We also evaluated all combinations of two architectural variants: First, we evaluated models in which we included the attention layer (LSTM+Attention) or simply used the final hidden state of the LSTM (LSTM) as a sentence representation. Second, since participants providing inference strength ratings also had access to the preceding conversational context, we also compared models that make predictions based only the target utterance with the some-NP and models that make predictions based on target utterances and the preceding conversational context. For the models using GloVe and ELMo, we prepended the conversational context to the target utterance to obtain a joint context and sentence embedding. For models using BERT, we made use of the fact that BERT had been trained to jointly embed two sentences or documents, and we obtained embeddings for the tokens in the target utterance by feeding the target utterance as the first document and the preceding context as the second document into the BERT encoder. For these models, we discarded the hidden states of the preceding context and only used the output of the BERT encoder for the tokens in the target utterance. Implementation details. We implemented the model in PyTorch BIBREF34. We trained the model using the Adam optimizer BIBREF35 with default parameters and a learning rate of 0.001, minimizing the mean squared error of the predicted ratings. In the no-context experiments, we truncated target utterances longer than 30 tokens, and in the experiments with context, we truncated the beginning of the preceding context such that the number of tokens did not exceed 150. Evaluation. We evaluated the model predictions in terms of their correlation $r$ with the human inference strength ratings. For the optimization of hyperparameters and architectural variants, we evaluated the model using 5-fold cross-validation. We then took the best set of parameters and trained a model on all the available training data and evaluated that model on the held-out data. ### Experiments ::: Tuning results
We find that the attention layer improves predictions; that contextual word embeddings lead to better results than the static GloVe embeddings; and that including the conversational context does not improve predictions (see Appendix SECREF8, for learning curves of all models, and Section SECREF6, for a discussion of the role of conversational context). Otherwise, the model is quite insensitive to hyperparameter settings: neither the dimension of the hidden LSTM states nor the dropout rate had considerable effects on the prediction accuracy. We do find, however, that there are differences depending on the BERT and ELMo layer that we use as word representations. We find that higher layers work better than lower layers, suggesting that word representations that are influenced by other utterance tokens are helpful for this task. Based on these optimization runs, we chose the model with attention that uses the BERT-large embeddings but no conversational context for the subsequent experiments and analyses. ### Experiments ::: Test results
Figure FIGREF26 shows the correlation between the best model according to the tuning runs (now trained on all training data) and the empirical ratings on the 408 held-out test items. As this plot shows, the model predictions fall within a close range of the empirical ratings for most of the items ($r=0.78$). Further, similarly as in the empirical data, there seem to be two clusters in the model predictions: one that includes lower ratings and one that includes higher ratings, corresponding to strong and weak scalar inferences, respectively. The only systematic deviation appears to be that the model does not predict any extreme ratings – almost all predictions are greater than 2 or less than 6, whereas the empirical ratings include some cases outside of this range. Overall, these results suggest that the model can learn to closely predict the strength of scalar inferences. However, this result by itself does not provide evidence that the model learned associations between linguistic cues and inference strength, since it could also be that, given the large number of parameters, the model learned spurious correlations independent of the empirically established cue-strength associations. To rule out the latter explanation, we probed the model's behavior in multiple ways, which we discuss next. ### Model behavior analyses
Regression analysis. As a first analysis, we investigated whether the neural network model predictions explain (some of) the variance explained by the linguistic factors that modulate inference strength. To this end, we used a slightly simplified Bayesian implementation of the linear mixed-effects model by degen2015investigating using the brms BIBREF36 and STAN BIBREF37 packages and compared this original model to an extended model that included the output of the above NN model as a predictor. For this comparison, we investigated whether the magnitude of a predictor in the original model significantly decreased in the extended model that included the NN predictions, based on the reasoning that if the NN predictions already explain the variance previously explained by these manually coded predictors, then the original predictor should explain no or less additional variance. We approximated the probability that the magnitude of the coefficient in the extended model including the NN predictor is smaller than the coefficient in the original model, $P(|\beta _i^{extended}| < |\beta _i^{original}|)$, by sampling values for each coefficient from the distributions of the original and the extended models and comparing the magnitude of the sampled coefficients. We repeated this process 1,000,000 times and treated the simulated proportions as approximate probabilities. An issue with this analysis is that estimating the regression model only on the items in the held-out test set yields very wide credible intervals for some of the predictors–in particular for some of the interactions–since the model infers these values from very little data. We therefore performed this (and all subsequent) analyses on the entire data, and obtained the NN predictions through 6-fold cross-validation, so that the NN model always made predictions on data that it had not seen during training. This did yield the same qualitative results as the analyses only performed on the held-out test items (see Appendix SECREF9) but it also provided us with narrower credible intervals that highlight the differences between the coefficient estimates of the two models. Figure FIGREF27 shows the estimates of the coefficients in the original model and the extended model. We find that the NN predictions explain some or all of the variance originally explained by many of the manually coded linguistic features: the estimated magnitude of the predictors for partitive, determiner strength, linguistic mention, subjecthood, modification, utterance length, and two of the interaction terms decreased in the extended model. These results suggest that the NN model indeed learned associations between linguistic features and inference strength rather than only explaining variance caused by individual items. This is particularly true for the grammatical and lexical features; we find that the NN predictor explains most of the variance originally explained by the partitive, subjecthood, and modification predictors. More surprisingly, the NN predictions also explain a lot of the variance originally explained by the determiner strength predictor, which was empirically determined by probing human interpretation and is not encoded explicitly in the surface form utterance. One potential explanation for this is that strong and weak some have different context distributions. For instance, weak some occurs in existential there constructions and with individual-level predicates, whereas strong some tends not to BIBREF23, BIBREF38, BIBREF39. Since pre-trained word embedding models capture a lot of distributional information, the NN model is presumably able to learn this association. ### Model behavior analyses ::: Attention weight analysis.
As a second type of analysis, we analyzed the attention weights that the model used for combining the token embeddings to a sentence embedding. Attention weight analyses have been successfully used for inspecting and debugging model decisions (e.g., BIBREF40, BIBREF41, BIBREF42, BIBREF43; but see BIBREF44, and BIBREF45, for critical discussions of this approach). Based on these results, we expected the model to attend more to tokens that are relevant for making predictions. Given that many of the hand-mined features that predict inference strength occur within or in the vicinity of the some-NP, we should therefore expect the model to attend most to the some-NP. To test this, we first explored whether the model attended on average more to some than to other tokens in the same position. Further, we exploited the fact that subjects generally occur at the beginning of a sentence. If the model attends to the vicinity of the some-NP, the average attention weights should be higher on early positions for utterances with a subject some-NP compared to utterances with a non-subject some-NP, and conversely for late utterance positions. We thus compared the average attention weights for each position in the utterance across utterances with subject versus non-subject some-NPs. To make sure that any effects were not only driven by the attention weight of the some-tokens, we set the attention weights of the token corresponding to some to 0 and re-normalized the attention weights for this analysis. Further, since the attention weights are dependent on the number of tokens in the utterance, it is crucial that the average utterance length across the two compared groups be matched. We addressed this by removing outliers and limiting our analysis to utterances up to length 30 (1,028 utterances), which incidentally equalized the number of tokens across the two groups. (While these exclusions resulted in tiny quantitative difference in the average attention weights, the qualitative patterns are not affected.) The left panel of Figure FIGREF30 shows the average attention weight by position for some versus other tokens. The model assigns much higher weight to some. The center panel of Figure FIGREF30 shows the average attention weight by position for subject vs. non-subject some-NP utterances. The attention weights are generally higher for tokens early in the utterance, but the attention weights of utterances with a subject some-NP are on average higher for tokens early in the utterance compared to utterances with the some-NP in non-subject positions. Both of these findings provide evidence that the model assigns high weight to the tokens within and surrounding the some-NP. In a more targeted analysis to assess whether the model learned to use the partitive cue, we examined whether the model assigned higher attention to the preposition of in partitive some-NPs compared to when of occurred elsewhere. As utterance length was again a potential confound, we conducted the analysis separately on the full set of utterances with raw attention weights and on a subset that included only utterances with at least two instances of of (128 utterances), in which we renormalized the weights of of-tokens to sum to 1. Results are shown in the right panel of Figure FIGREF30. The attention weights were higher for of tokens in partitive some-NPs, suggesting that the model learned an association between partitive of in some-NPs and inference strength. ### Model behavior analyses ::: Minimal pair analysis.
As a final analysis, we constructed artificial minimal pairs that differed along several factors of interest and compared the model predictions. Such methods have been recently used to probe what kind of syntactic dependencies different types of recurrent neural network language models are capable of encoding (e.g., BIBREF12, BIBREF13, BIBREF47, BIBREF48, BIBREF14, BIBREF15), and also allow us to probe whether the model is sensitive to controlled changes in the input. We constructed a set of 25 initial sentences with some-NPs. For each sentence, we created 32 variants that differed in the following four properties of the some-NP: subjecthood, partitive, pre-nominal modification, and post-nominal modification. For the latter three features, we either included or excluded of the or the modifier, respectively. To manipulate subjecthood of the some-NP, we created variants in which some was either the determiner in the subject NP as in (UNKREF36a) or in the object-NP as in (UNKREF36b). We also created passive versions of each of these variants (UNKREF36c-d). Each set of sentences included a unique main verb, a unique pair of NPs, and unique modifiers. The full list of sentences can be found in Appendix SECREF10. Some (of the) (organic) farmers (in the mountains) milked the brown goats who graze on the meadows. The organic farmers in the mountains milked some (of the) (brown) goats (who graze on the meadows). The brown goats who graze on the meadows were milked by some (of the) (organic) farmers (in the mountains). Some (of the) (brown) goats (who graze on the meadows) were milked by the organic farmers in the mountains. Figure FIGREF41 shows the model predictions for the manually constructed sentences grouped by the presence of a partitive construction, the grammatical function of the some-NP, and the presence of a modifier. As in the natural dataset from BIBREF22, sentences with a partitive received higher predicted ratings than sentences without a partitive; sentences with subject some-NPs received higher predicted ratings than sentences with non-subject some-NPs; and sentences with a modified head noun in the some-NP received lower predictions than sentences with an unmodified some-NP. All these results provide additional evidence that the model learned the correct associations. This is particularly remarkable considering the train-test mismatch: the model was trained on noisy transcripts of spoken language that contained many disfluencies and repairs, and was subsequently tested on clean written sentences. ### Context, revisited
In the tuning experiments above, we found that including the preceding conversational context in the input to the model did not improve or lowered prediction accuracy. At the same time, we found that the model is capable of making accurate predictions in most cases without taking the preceding context into account. Taken together, these results suggest either that the conversational context is not necessary and one can draw inferences from the target utterance alone, or that the model does not make adequate use of the preceding context. BIBREF22 did not systematically investigate whether the preceding conversational context was used by participants judging inference strength. To assess the extent to which the preceding context in this dataset affects inference strength, we re-ran her experiment, but without presenting participants with the preceding conversational context. If the context is irrelevant for drawing inferences, then mean inference strength ratings should be very similar across the two experiments, suggesting the model may have rightly learned to not utilize the context. If the presence of context affects inference strength, ratings should differ across experiments, suggesting that the model's method of integrating context is ill-suited to the task. The new, no-context ratings correlated with the original ratings ($r=0.68$, see Appendix SECREF11) but were overall more concentrated towards the center of the scale, suggesting that in many cases, participants who lacked information about the conversational context were unsure about the strength of the scalar inference. Since the original dataset exhibited more of a bi-modal distribution with fewer ratings at the center of the scale, this suggests that the broader conversational context contains important cues to scalar inferences. For our model, these results suggest that the representation of the conversational context is inadequate, which highlights the need for more sophisticated representations of linguistic contexts beyond the target utterance. We further find that the model trained on the original dataset is worse at predicting the no-context ratings ($r=0.66$) than the original ratings ($r=0.78$), which is not surprising considering the imperfect correlation between ratings across experiments, but also provides additional evidence that participants indeed behaved differently in the two experiments. ### Conclusion and future work
We showed that neural network-based sentence encoders are capable of harnessing the linguistic signal to learn to predict human inference strength ratings from some to not all with high accuracy. Further, several model behavior analyses provided consistent evidence that the model learned associations between previously established linguistic features and the strength of scalar inferences. Taken together, these results suggest that it is possible to learn associations between linguistic features and scalar inferences from statistical input consisting of a relatively small set of utterances. In an analysis of the contribution of the conversational context, we found that humans make use of the preceding context whereas the models we considered failed to do so adequately. Considering the importance of context in drawing both scalar and other inferences in communication BIBREF0, BIBREF51, BIBREF52, BIBREF53, BIBREF54, BIBREF6, BIBREF7, the development of appropriate representations of larger context is an exciting avenue for future research. One further interesting line of research would be to extend this work to other pragmatic inferences. Recent experimental work has shown that inference strength is variable across scale and inference type BIBREF55, BIBREF56. We treated some as a case study in this work, but none of our modeling decisions are specific to some. It would be straightforward to train similar models for other types of inferences. Lastly, the fact that the attention weights provided insights into the model's decisions suggests possibilities for using neural network models for developing more precise theories of pragmatic language use. Our goal here was to investigate whether neural networks can learn associations for already established linguistic cues but it would be equally interesting to investigate whether such models could be used to discover new cues, which could then be verified in experimental and corpus work, potentially providing a novel model-driven approach to experimental and formal pragmatics. ### Hyperparameter tuning
Figure FIGREF44 shows the learning curves averaged over the 5 cross-validation tuning runs for models using different word embeddings. As these plots show, the attention layer improves predictions; contextual word embeddings lead to better results than the static GloVe embeddings; and including the conversational context does not improve predictions and in some cases even lowers prediction accuracy. ### Regression analysis on held-out test data
Figure FIGREF45 shows the estimates of the predictors in the original and extended Bayesian mixed-effects models estimated only on the held-out test data. We find the same qualitative effects as in Figure FIGREF27, but since these models were estimated on much less data (only 408 items), there is a lot of uncertainty in the estimates and therefore quantitative comparisons between the coefficients of the different models are less informative. ### List of manually constructed sentences
Tables TABREF46 and TABREF47 show the 25 manually created sentences for the analyses described in the minimal pairs analysis in Section SECREF5. As described in the main text, we created 16 variants of the sentence with the some-NP in subject position (sentences in the left column), and 16 variants of the sentence with the some-NP in object position (sentences in the right column), yielding in total 800 examples. ### Results from no-context experiment
Figure FIGREF48 shows the correlation between the mean inference strength ratings for each item in the experiment from BIBREF22 and the mean strength ratings from the new no-context experiment, discussed in Section SECREF6. Figure 1: Model architecture. Figure 2: Correlation between empirical ratings and predictions of the BERT-LARGE LSTM+ATTENTION model on held-out test items. Figure 3: Maximum a posteriori estimates and 95%-credible intervals of coefficients for original and extended Bayesian mixed-effects regression models predicting the inference strength ratings. */**/*** indicate that the probability of the coefficient of the original model having a larger magnitude than the coefficient of the extended model is less than 0.05, 0.01, and 0.001, respectively. Figure 4: Left: Average attention weights at each token position for some and other tokens. Center: Average attention weights at each token position for utterances with subject and non-subject some-NPs. Right: Average attention weights of of -tokens in partitive some-NPs and weights of other of -tokens. In the normalized cases, we take only the utterances with multiple of -tokens into account and re-normalize the attention weights across all of -tokens in one utterance. Error bars indicate 95% bootstrapped confidence intervals. Figure 5: Average model predictions on manually constructed sentences, grouped by presence of partitives, by grammatical function of the some-NP, and by presence of nominal modifiers. Semi-transparent dots show predictions on individual sentences. Figure 6: Correlation between each model’s predictions on valuation set and empirical means, by training epoch. Figure 7: Maximum a posteriori estimates and 95%-credible intervals of coefficients for original and extended Bayesian mixed-effects regression models predicting the inference strength ratings on the held-out test set. */**/*** indicate that the probability of the coefficient of the original model having a larger magnitude than the coefficient of the extended model is less than 0.05, 0.01, and 0.001, respectively. Table 1: Manually constructed sentences used in the minimal pair analyses. Sentences in the left column have a some-NP in subject position; sentences on the right have a some-NP object position. Table 2: Manually constructed sentences used in the minimal pair analyses (continued). Figure 8: Mean inference strength ratings for items without context (new) against items with context (original), r = .68. | the annotated dataset reported by degen2015investigating, a dataset of the utterances from the Switchboard corpus of telephone dialogues BIBREF21 that contain the word some |
What are lyrical topics present in the metal genre? | ### Introduction
As audio and text features provide complementary layers of information on songs, a combination of both data types has been shown to improve the automatic classification of high-level attributes in music such as genre, mood and emotion BIBREF0, BIBREF1, BIBREF2, BIBREF3. Multi-modal approaches interlinking these features offer insights into possible relations between lyrical and musical information (see BIBREF4, BIBREF5, BIBREF6). In the case of metal music, sound dimensions like loudness, distortion and particularly hardness (or heaviness) play an essential role in defining the sound of this genre BIBREF7, BIBREF8, BIBREF9, BIBREF10. Specific subgenres – especially doom metal, gothic metal and black metal – are further associated with a sound that is often described as dark or gloomy BIBREF11, BIBREF12. These characteristics are typically not limited to the acoustic and musical level. In a research strand that has so far been generally treated separately from the audio dimensions, lyrics from the metal genre have come under relatively close scrutiny (cf. BIBREF13). Topics typically ascribed to metal lyrics include sadness, death, freedom, nature, occultism or unpleasant/disgusting objects and are overall characterized as harsh, gloomy, dystopian, or satanic BIBREF14, BIBREF13, BIBREF15, BIBREF16, BIBREF17. Until now, investigations on metal lyrics were limited to individual cases or relatively small corpora – with a maximum of 1,152 songs in BIBREF17. Besides this, the relation between the musical and the textual domain has not yet been explored. Therefore, we examine a large corpus of metal song lyrics, addressing the following questions: Which topics are present within the corpus of metal lyrics? Is there a connection between characteristic musical dimensions like hardness and darkness and certain topics occurring within the textual domain? ### Methodology
In our sequential research design, the distribution of textual topics within the corpus was analyzed using latent Dirichlet allocation (LDA). This resulted in a topic model, which was used for a probabilistic assignment of topics to each of the song documents. Additionally, for a subset of these songs, audio features were extracted using models for high-level music dimensions. The use of automatic models for the extraction of both text as well as musical features allows for scalability as it enables a large corpus to be studied without depending on the process of manual annotation for each of the songs. The resulting feature vectors were then subjected to a correlation analysis. Figure FIGREF6 outlines the sequence of the steps taken in processing the data. The individual steps are explained in the following subsections. ### Methodology ::: Text Corpus Creation and Cleaning
For gathering the data corpus, a web crawler was programmed using the Python packages Requests and BeautifulSoup. In total, 152,916 metal music lyrics were extracted from www.darklyrics.com. Using Python’s langdetect package, all non-English texts were excluded. With the help of regular expressions, the texts were scanned for tokens indicating meta-information, which is not part of the actual lyrics. To this end, a list of stopwords referring to musical instruments or the production process (e.g. ‘recorded’, ‘mixed’, ‘arrangement by’, ‘band photos’) was defined in addition to common stopwords. After these cleaning procedures, 124,288 texts remained in the subsample. For text normalization, stemming and lemmatization were applied as further preprocessing steps. ### Methodology ::: Topic Modelling via Latent Dirichlet Allocation
We performed a LDA BIBREF18 on the remaining subsample to construct a probabilistic topic model. The LDA models were created by using the Python library Gensim BIBREF19. The lyrics were first converted to a bag-of-words format, and standard weighting of terms provided by the Gensim package was applied. Log perplexity BIBREF20 and log UMass coherence BIBREF21 were calculated as goodness-of-fit measures evaluating topic models ranging from 10 to 100 topics. Considering these performance measures as well as qualitative interpretability of the resulting topic models, we chose a topic model including 20 topics – an approach comparable with BIBREF22. We then examined the most salient and most typical words for each topic. Moreover, we used the ldavis package to analyze the structure of the resulting topic space BIBREF23. In order to do so, the Jensen-Shannon divergence between topics was calculated in a first step. In a second step, we applied multidimensional scaling (MDS) to project the inter-topic distances onto a two-dimensional plane. MDS is based on the idea of calculating dissimilarities between pairs of items of an input matrix while minimizing the strain function BIBREF24. In this case, the closer the topics are located to one another on the two-dimensional plane, the more they share salient terms and the more likely a combination of these topics appear in a song. ### Methodology ::: High-Level Audio Feature Extraction
The high-level audio feature models used had been constructed in previous examinations BIBREF25, BIBREF26. In those music perception studies, ratings were obtained for 212 music stimuli in an online listening experiment by 40 raters. 2 Based on this ground truth, prediction models for the automatic extraction of high-level music dimensions – including the concepts of perceived hardness/heaviness and darkness/gloominess in music – had been trained using machine learning methods. In a second step, the model obtained for hardness had been evaluated using further listening experiments on a new unseen set of audio stimuli BIBREF26. The model has been refined against this backdrop, resulting in an $R^2$ value of 0.80 for hardness/heaviness and 0.60 for darkness/gloominess using five-fold cross-validation. The resulting models embedded features implemented in LibROSA BIBREF27, Essentia BIBREF28 as well as the timbral models developed as part of the AudioCommons project BIBREF29. ### Methodology ::: Investigating the Connection between Audio and Text Features
Finally, we drew a random sample of 503 songs and used Spearman's $\rho $ to identify correlations between the topics retrieved and the audio dimensions obtained by the high-level audio feature models. We opted for Spearman’s $\rho $ since it does not assume normal distribution of the data, is less prone to outliers and zero-inflation than Pearson’s $r$. Bonferroni correction was applied in order to account for multiple-testing. ### Results ::: Textual Topics
Table TABREF10 displays the twenty resulting topics found within the text corpus using LDA. The topics are numbered in descending order according to their prevalence (weight) in the text corpus. For each topic, a qualitative interpretation is given along with the 10 most salient terms. The salient terms of the first topic – and in parts also the second – appear relatively generic, as terms like e.g. ‘know’, ‘never’, and ‘time’ occur in many contexts. However, the majority of the remaining topics reveal distinct lyrical themes described as being characteristic for the metal genre. ‘Religion & satanism’ (topic #5) and descriptions of ‘brutal death’ (topic #7) can be considered as being typical for black metal and death metal respectively, whereas ‘battle’ (topic #6), ‘landscape & journey’ (topic #11), ‘struggle for freedom’ (topic #12), and ‘dystopia’ (topic #15), are associated with power metal and other metal subgenres. 2 This is highlighted in detail in Figure FIGREF11. Here, the topic distributions for two exemplary bands contained within the sample are presented. For these heat maps, data has been aggregated over individual songs showing the topic distribution at the level of albums over a band’s history. The examples chosen illustrate the dependence between textual topics and musical subgenres. For the band Manowar, which is associated with the genre of heavy metal, power metal or true metal, a prevalence of topic #6 (‘battle’) can be observed, while a distinctive prevalence of topic #7 (‘brutal death’) becomes apparent for Cannibal Corpse – a band belonging to the subgenre of death metal. Within the topic configuration obtained via multidimensional scaling (see Figure FIGREF12), two latent dimensions can be identified. The first dimension (PC1) distinguishes topics with more common wordings on the right hand side from topics with less common wording on the left hand side. This also correlates with the weight of the topics within the corpus. The second dimension (PC2) is characterized by an contrast between transcendent and sinister topics dealing with occultism, metaphysics, satanism, darkness, and mourning (#9, #3, .#5, #13, and #16) at the top and comparatively shallow content dealing with personal life and Rock’n’Roll lifestyle using a rather mundane or vulgar vocabulary (#1, #8, and #19) at the bottom. This contrast can be interpreted as ‘otherworldliness / individual-transcending narratives’ vs. ‘worldliness / personal life’. ### Results ::: Correlations with Musical Dimensions
In the final step of our analysis, we calculated the association between the twenty topics discussed above and the two high-level audio features hardness and darkness using Spearman’s $\rho $. The results are visualized in Figure FIGREF13 and the $\rho $ values listed in table TABREF10. Significant positive associations can be observed between musical hardness and the topics ‘brutal death’, ‘dystopia’, ‘archaisms & occultism’, ‘religion & satanism’, and ‘battle’, while it is negatively linked to relatively mundane topics concerning ‘personal life’ and ‘love & romance’. The situation is similar for dark/gloomy sounding music, which in turn is specifically related to themes such as ‘dystopia’ and ‘(psychological) madness’. Overall, the strength of the associations is moderate at best, with a tendency towards higher associations for hardness than darkness. The strongest association exists between hardness and the topic ‘brutal death’ ($\rho = 0.267$, $p < 0.01$). ### Conclusion and Outlook
Applying the example of metal music, our work examined the textual topics found in song lyrics and investigated the association between these topics and high-level music features. By using LDA and MDS in order to explore prevalent topics and the topic space, typical text topics identified in qualitative analyses could be confirmed and objectified based on a large text corpus. These include e.g. satanism, dystopia or disgusting objects. It was shown that musical hardness is particularly associated with harsh topics like ‘brutal death’ and ‘dystopia’, while it is negatively linked to relatively mundane topics concerning personal life and love. We expect that even stronger correlations could be found for metal-specific topics when including more genres covering a wider range of hardness/darkness values. Therefore, we suggest transferring the method to a sample including multiple genres. Moreover, an integration with metadata such as genre information would allow for the testing of associations between topics, genres and high-level audio features. This could help to better understand the role of different domains in an overall perception of genre-defining attributes such as hardness. Figure 1: Processing steps of the approach illustrating the parallel analysis of text and audio features Table 1: Overview of the resulting topics found within the corpus of metal lyrics (n = 124,288) and their correlation to the dimensions hardness and darkness obtained from the audio signal (see section 3.2) Figure 2: Comparison of the topic distributions for all included albums by the bands Manowar and Cannibal Corpse showing a prevalence of the topics ‘battle’ and ‘brutal death’ respectively Figure 3: Topic configuration obtained via multidimensional scaling. The radius of the circles is proportional to the percentage of tokens covered by the topics (topic weight). Figure 4: Correlations between lyrical topics and the musical dimensions hardness and darkness; ∗: p < 0.05, ∗∗: p < 0.00125 (Bonferroni-corrected significance level) | Table TABREF10 displays the twenty resulting topics |
What multilingual word representations are used? | ### Motivations
Figurative language makes use of figures of speech to convey non-literal meaning BIBREF0, BIBREF1. It encompasses a variety of phenomena, including metaphor, humor, and irony. We focus here on irony and uses it as an umbrella term that covers satire, parody and sarcasm. Irony detection (ID) has gained relevance recently, due to its importance to extract information from texts. For example, to go beyond the literal matches of user queries, Veale enriched information retrieval with new operators to enable the non-literal retrieval of creative expressions BIBREF2. Also, the performances of sentiment analysis systems drastically decrease when applied to ironic texts BIBREF3, BIBREF4. Most related work concern English BIBREF5, BIBREF6 with some efforts in French BIBREF7, Portuguese BIBREF8, Italian BIBREF9, Dutch BIBREF10, Hindi BIBREF11, Spanish variants BIBREF12 and Arabic BIBREF13, BIBREF14. Bilingual ID with one model per language has also been explored, like English-Czech BIBREF15 and English-Chinese BIBREF16, but not within a cross-lingual perspective. In social media, such as Twitter, specific hashtags (#irony, #sarcasm) are often used as gold labels to detect irony in a supervised learning setting. Although recent studies pointed out the issue of false-alarm hashtags in self-labeled data BIBREF17, ID via hashtag filtering provides researchers positive examples with high precision. On the other hand, systems are not able to detect irony in languages where such filtering is not always possible. Multilingual prediction (either relying on machine translation or multilingual embedding methods) is a common solution to tackle under-resourced languages BIBREF18, BIBREF19. While multilinguality has been widely investigated in information retrieval BIBREF20, BIBREF21 and several NLP tasks (e.g., sentiment analysis BIBREF22, BIBREF23 and named entity recognition BIBREF24), no one explored it for irony. We aim here to bridge the gap by tackling ID in tweets from both multilingual (French, English and Arabic) and multicultural perspectives (Indo-European languages whose speakers share quite the same cultural background vs. less culturally close languages). Our approach does not rely either on machine translation or parallel corpora (which are not always available), but rather builds on previous corpus-based studies that show that irony is a universal phenomenon and many languages share similar irony devices. For example, Karoui et. al BIBREF25 concluded that their multi-layer annotated schema, initially used to annotate French tweets, is portable to English and Italian, observing relatively the same tendencies in terms of irony categories and markers. Similarly, Chakhachiro BIBREF26 studies irony in English and Arabic, and shows that both languages share several similarities in the rhetorical (e.g., overstatement), grammatical (e.g., redundancy) and lexical (e.g., synonymy) usage of irony devices. The next step now is to show to what extent these observations are still valid from a computational point of view. Our contributions are: A new freely available corpus of Arabic tweets manually annotated for irony detection. Monolingual ID: We propose both feature-based models (relying on language-dependent and language-independent features) and neural models to measure to what extent ID is language dependent. Cross-lingual ID: We experiment using cross-lingual word representation by training on one language and testing on another one to measure how the proposed models are culture-dependent. Our results are encouraging and open the door to ID in languages that lack of annotated data for irony. ### Data
Arabic dataset (Ar=$11,225$ tweets). Our starting point was the corpus built by BIBREF13 that we extended to different political issues and events related to the Middle East and Maghreb that hold during the years 2011 to 2018. Tweets were collected using a set of predefined keywords (which targeted specific political figures or events) and containing or not Arabic ironic hashtags (سخرية>#, مسخرة>#, تهكم>#, استهزاء>#) . The collection process resulted in a set of $6,809$ ironic tweets ($I$) vs. $15,509$ non ironic ($NI$) written using standard (formal) and different Arabic language varieties: Egypt, Gulf, Levantine, and Maghrebi dialects. To investigate the validity of using the original tweets labels, a sample of $3,000$ $I$ and $3,000$ $NI$ was manually annotated by two Arabic native speakers which resulted in $2,636$ $I$ vs. $2,876$ $NI$. The inter-annotator agreement using Cohen's Kappa was $0.76$, while the agreement score between the annotators' labels and the original labels was $0.6$. Agreements being relatively good knowing the difficulty of the task, we sampled $5,713$ instances from the original unlabeled dataset to our manually labeled part. The added tweets have been manually checked to remove duplicates, very short tweets and tweets that depend on external links, images or videos to understand their meaning. French dataset (Fr=$7,307$ tweets). We rely on the corpus used for the DEFT 2017 French shared task on irony BIBREF3 which consists of tweets relative to a set of topics discussed in the media between 2014 and 2016 and contains topic keywords and/or French irony hashtags (#ironie, #sarcasme). Tweets have been annotated by three annotators (after removing the original labels) with a reported Cohen's Kappa of $0.69$. English dataset (En=$11,225$ tweets). We use the corpus built by BIBREF15 which consists of $100,000$ tweets collected using the hashtag #sarcasm. It was used as benchmark in several works BIBREF27, BIBREF28. We sliced a subset of approximately $11,200$ tweets to match the sizes of the other languages' datasets. Table TABREF6 shows the tweet distribution in all corpora. Across the three languages, we keep a similar number of instances for train and test sets to have fair cross-lingual experiments as well (see Section SECREF4). Also, for French, we use the original dataset without any modification, keeping the same number of records for train and test to better compare with state-of-the-art results. For the classes distribution (ironic vs. non ironic), we do not choose a specific ratio but we use the resulted distribution from the random shuffling process. ### Monolingual Irony Detection
It is important to note that our aim is not to outperform state-of-the-art models in monolingual ID but to investigate which of the monolingual architectures (neural or feature-based) can achieve comparable results with existing systems. The result can show which kind of features works better in the monolingual settings and can be employed to detect irony in a multilingual setting. In addition, it can show us to what extend ID is language dependent by comparing their results to multilingual results. Two models have been built, as explained below. Prior to learning, basic preprocessing steps were performed for each language (e.g., removing foreign characters, ironic hashtags, mentions, and URLs). Feature-based models. We used state-of-the-art features that have shown to be useful in ID: some of them are language-independent (e.g., punctuation marks, positive and negative emoticons, quotations, personal pronouns, tweet's length, named entities) while others are language-dependent relying on dedicated lexicons (e.g., negation, opinion lexicons, opposition words). Several classical machine learning classifiers were tested with several feature combinations, among them Random Forest (RF) achieved the best result with all features. Neural model with monolingual embeddings. We used Convolutional Neural Network (CNN) network whose structure is similar to the one proposed by BIBREF29. For the embeddings, we relied on $AraVec$ BIBREF30 for Arabic, FastText BIBREF31 for French, and Word2vec Google News BIBREF32 for English . For the three languages, the size of the embeddings is 300 and the embeddings were fine-tuned during the training process. The CNN network was tuned with 20% of the training corpus using the $Hyperopt$ library. Results. Table TABREF9 shows the results obtained when using train-test configurations for each language. For English, our results, in terms of macro F-score ($F$), were not comparable to those of BIBREF15, BIBREF33, as we used 11% of the original dataset. For French, our scores are in line with those reported in state of the art (cf. best system in the irony shared task achieved $F=78.3$ BIBREF3). They outperform those obtained for Arabic ($A=71.7$) BIBREF13 and are comparable to those recently reported in the irony detection shared task in Arabic tweets BIBREF14, BIBREF34 ($F=84.4$). Overall, the results show that semantic-based information captured by the embedding space are more productive comparing to standard surface and lexicon-based features. ### Cross-lingual Irony Detection
We use the previous CNN architecture with bilingual embedding and the RF model with surface features (e.g., use of personal pronoun, presence of interjections, emoticon or specific punctuation) to verify which pair of the three languages: (a) has similar ironic pragmatic devices, and (b) uses similar text-based pattern in the narrative of the ironic tweets. As continuous word embedding spaces exhibit similar structures across (even distant) languages BIBREF35, we use a multilingual word representation which aims to learn a linear mapping from a source to a target embedding space. Many methods have been proposed to learn this mapping such as parallel data supervision and bilingual dictionaries BIBREF35 or unsupervised methods relying on monolingual corpora BIBREF36, BIBREF37, BIBREF38. For our experiments, we use Conneau et al 's approach as it showed superior results with respect to the literature BIBREF36. We perform several experiments by training on one language ($lang_1$) and testing on another one ($lang_2$) (henceforth $lang_1\rightarrow lang_2$). We get 6 configurations, plus two others to evaluate how irony devices are expressed cross-culturally, i.e. in European vs. non European languages. In each experiment, we took 20% from the training to validate the model before the testing process. Table TABREF11 presents the results. From a semantic perspective, despite the language and cultural differences between Arabic and French languages, CNN results show a high performance comparing to the other languages pairs when we train on each of these two languages and test on the other one. Similarly, for the French and English pair, but when we train on French they are quite lower. We have a similar case when we train on Arabic and test on English. We can justify that by, the language presentation of the Arabic and French tweets are quite informal and have many dialect words that may not exist in the pretrained embeddings we used comparing to the English ones (lower embeddings coverage ratio), which become harder for the CNN to learn a clear semantic pattern. Another point is the presence of Arabic dialects, where some dialect words may not exist in the multilingual pretrained embedding model that we used. On the other hand, from the text-based perspective, the results show that the text-based features can help in the case when the semantic aspect shows weak detection; this is the case for the $Ar\longrightarrow En$ configuration. It is worthy to mention that the highest result we get in this experiment is from the En$\rightarrow $Fr pair, as both languages use Latin characters. Finally, when investigating the relatedness between European vs. non European languages (cf. (En/Fr)$\rightarrow $Ar), we obtain similar results than those obtained in the monolingual experiment (macro F-score 62.4 vs. 68.0) and best results are achieved by Ar $\rightarrow $(En/Fr). This shows that there are pragmatic devices in common between both sides and, in a similar way, similar text-based patterns in the narrative way of the ironic tweets. ### Discussions and Conclusion
This paper proposes the first multilingual ID in tweets. We show that simple monolingual architectures (either neural or feature-based) trained separately on each language can be successfully used in a multilingual setting providing a cross-lingual word representation or basic surface features. Our monolingual results are comparable to state of the art for the three languages. The CNN architecture trained on cross-lingual word representation shows that irony has a certain similarity between the languages we targeted despite the cultural differences which confirm that irony is a universal phenomena, as already shown in previous linguistic studies BIBREF39, BIBREF25, BIBREF40. The manual analysis of the common misclassified tweets across the languages in the multilingual setup, shows that classification errors are due to three main factors. (1) First, the absence of context where writers did not provide sufficient information to capture the ironic sense even in the monolingual setting, as in نبدا تاني يسقط يسقط حسني مبارك !! > (Let's start again, get off get off Mubarak!!) where the writer mocks the Egyptian revolution, as the actual president "Sisi" is viewed as Mubarak's fellows. (2) Second, the presence of out of vocabulary (OOV) terms because of the weak coverage of the mutlilingual embeddings which make the system fails to generalize when the OOV set of unseen words is large during the training process. We found tweets in all the three languages written in a very informal way, where some characters of the words were deleted, duplicated or written phonetically (e.g phat instead of fat). (3) Another important issue is the difficulty to deal with the Arabic language. Arabic tweets are often characterized by non-diacritised texts, a large variations of unstandardized dialectal Arabic (recall that our dataset has 4 main varieties, namely Egypt, Gulf, Levantine, and Maghrebi), presence of transliterated words (e.g. the word table becomes طابلة> (tabla)), and finally linguistic code switching between Modern Standard Arabic and several dialects, and between Arabic and other languages like English and French. We found some tweets contain only words from one of the varieties and most of these words do not exist in the Arabic embeddings model. For example in مبارك بقاله كام يوم مامتش .. هو عيان ولاه ايه #مصر > (Since many days Mubarak didn't die .. is he sick or what? #Egypt), only the words يوم> (day), مبارك> (Mubarak), and هو> (he) exist in the embeddings. Clearly, considering only these three available words, we are not able to understand the context or the ironic meaning of the tweet. To conclude, our multilingual experiments confirmed that the door is open towards multilingual approaches for ID. Furthermore, our results showed that ID can be applied to languages that lack of annotated data. Our next step is to experiment with other languages such as Hindi and Italian. ### Acknowledgment
The work of Paolo Rosso was partially funded by the Spanish MICINN under the research project MISMIS-FAKEnHATE (PGC2018-096212-B-C31). Table 1. Tweet distribution in all corpora. Table 2. Results of the monolingual experiments (in percentage) in terms of accuracy (A), precision (P), recall (R), and macro F-score (F). Table 3. Results of the cross-lingual experiments. | a multilingual word representation which aims to learn a linear mapping from a source to a target embedding space |
What are the components of the multilingual framework? | ### Introduction
Sentiment analysis is a crucial task in opinion mining field where the goal is to extract opinions, emotions, or attitudes to different entities (person, objects, news, among others). Clearly, this task is of interest for all languages; however, there exists a significant gap between English state-of-the-art methods and other languages. It is expected that some researchers decided to test the straightforward approach which consists in, first, translating the messages to English, and, then, use a high performing English sentiment classifier (for instance, see BIBREF0 and BIBREF1 ) instead of creating a sentiment classifier optimized for a given language. However, the advantages of a properly tuned sentiment classifier have been studied for different languages (for instance, see BIBREF2 , BIBREF3 , BIBREF4 , BIBREF5 ). This manuscript focuses on the particular case of multilingual sentiment analysis of short informal texts such as Twitter messages. Our aim is to provide an easy-to-use tool to create sentiment classifiers based on supervised learning (i.e., labeled dataset) where the classifier should be competitive to those sentiment classifiers carefully tuned by some given languages. Furthermore, our second contribution is to create a well-performing baseline to compare new sentiment classifiers in a broad range of languages or to bootstrap new sentiment analysis systems. Our approach is based on selecting the best text-transforming techniques that optimize some performance measures where the chosen techniques are robust to typical writing errors. In this context, we propose a robust multilingual sentiment analysis method, tested in eight different languages: Spanish, English, Italian, Arabic, German, Portuguese, Russian and Swedish. We compare our approach ranking in three international contests: TASS'15, SemEval'15-16 and SENTIPOLC'14, for Spanish, English and Italian respectively; the remaining languages are compared directly with the results reported in the literature. The experimental results locate our approach in good positions for all considered competitions; and excellent results in the other five languages tested. Finally, even when our method is almost cross-language, it can be extended to take advantage of language dependencies; we also provide experimental evidence of the advantages of using these language-dependent techniques. The rest of the manuscript is organized as follows. Section SECREF2 describes our proposed Sentiment Analysis method. Section SECREF3 describes the datasets and contests used to test our approach; whereas, the experimental results, and, the discussion are presented on Section SECREF4 . Finally, Section SECREF5 concludes. ### Our Approach: Multilingual Polarity Classification
We propose a method for multilingual polarity classification that can serve as a baseline as well as a framework to build more complex sentiment analysis systems due to its simplicity and availability as an open source software. As we mentioned, this baseline algorithm for multilingual Sentiment Analysis (B4MSA) was designed with the purpose of being multilingual and easy to implement. B4MSA is not a naïve baseline which is experimentally proved by evaluating it on several international competitions. In a nutshell, B4MSA starts by applying text-transformations to the messages, then transformed text is represented in a vector space model (see Subsection SECREF13 ), and finally, a Support Vector Machine (with linear kernel) is used as the classifier. B4MSA uses a number of text transformations that are categorized in cross-language features (see Subsection SECREF3 ) and language dependent features (see Subsection SECREF9 ). It is important to note that, all the text-transformations considered are either simple to implement or there is a well-known library (e.g. BIBREF6 , BIBREF7 ) to use them. It is important to note that to maintain the cross-language property, we limit ourselves to not use additional knowledge, this include knowledge from affective lexicons or models based on distributional semantics. To obtain the best performance, one needs to select those text-transformations that work best for a particular dataset, therefore, B4MSA uses a simple random search and hill-climbing (see Subsection SECREF14 ) in space of text-transformations to free the user from this delicate and time-consuming task. Before going into the details of each text-transformation, Table TABREF2 gives a summary of the text-transformations used as well as their parameters associated. ### Cross-language Features
We defined cross-language features as a set of features that could be applied in most similar languages, not only related language families such as Germanic languages (English, German, etc.), Romance languages (Spanish, Italian, etc.), among others; but also similar surface features such as punctuation, diacritics, symbol duplication, case sensitivity, etc. Later, the combination of these features will be explored to find the best configuration for a given classifier. Generally, Twitter messages are full of slang, misspelling, typographical and grammatical errors among others; in order to tackle these aspects we consider different parameters to study this effect. The following points are the parameters to be considered as spelling features. Punctuation (del-punc) considers the use of symbols such as question mark, period, exclamation point, commas, among other spelling marks. Diacritic symbols (del-diac) are commonly used in languages such as Spanish, Italian, Russian, etc., and its wrong usage is one of the main sources of orthographic errors in informal texts; this parameter considers the use or absence of diacritical marks. Symbol reduction (del-d1), usually, twitter messages use repeated characters to emphasize parts of the word to attract user's attention. This aspect makes the vocabulary explodes. We applied the strategy of replacing the repeated symbols by one occurrence of the symbol. Case sensitivity (lc) considers letters to be normalized in lowercase or to keep the original source; the aim is to cut the words that are the same in uppercase and lowercase. We classified around 500 most popular emoticons, included text emoticons, and the whole set of unicode emoticons (around INLINEFORM0 ) defined by BIBREF8 into three classes: positive, negative and neutral, which are grouped under its corresponding polarity word defined by the class name. Table TABREF6 shows an excerpt of the dictionary that maps emoticons to their corresponding polarity class. N-words (word sequences) are widely used in many NLP tasks, and they have also been used in Sentiment Analysis BIBREF9 and BIBREF10 . To compute the N-words, the text is tokenized and N-words are calculated from tokens. For example, let INLINEFORM0 be the text, so its 1-words (unigrams) are each word alone, and its 2-words (bigrams) set are the sequences of two words, the set ( INLINEFORM1 ), and so on. INLINEFORM2 = {the lights, lights and, and shadows, shadows of, of your, your future}, so, given text of size INLINEFORM3 words, we obtain a set containing at most INLINEFORM4 elements. Generally, N-words are used up to 2 or 3-words because it is uncommon to find, between texts, good matches of word sequences greater than three or four words BIBREF11 . In addition to the traditional N-words representation, we represent the resulting text as q-grams. A q-grams is an agnostic language transformation that consists in representing a document by all its substring of length INLINEFORM0 . For example, let INLINEFORM1 be the text, its 3-grams set are INLINEFORM2 so, given text of size INLINEFORM0 characters, we obtain a set with at most INLINEFORM1 elements. Notice that this transformation handles white-spaces as part of the text. Since there will be q-grams connecting words, in some sense, applying q-grams to the entire text can capture part of the syntactic and contextual information in the sentence. The rationale of q-grams is also to tackle misspelled sentences from the approximate pattern matching perspective BIBREF12 . ### Language Dependent Features
The following features are language dependent because they use specific information from the language concerned. Usually, the use of stopwords, stemming and negations are traditionally used in Sentiment Analysis. The users of this approach could add other features such as part of speech, affective lexicons, etc. to improve the performance BIBREF13 . In many languages, there is a set of extremely common words such as determiners or conjunctions ( INLINEFORM0 or INLINEFORM1 ) which help to build sentences but do not carry any meaning for themselves. These words are known as Stopwords, and they are removed from text before any attempt to classify them. Generally, a stopword list is built using the most frequent terms from a huge document collection. We used the Spanish, English and Italian stopword lists included in the NLTK Python package BIBREF6 in order to identify them. Stemming is a well-known heuristic process in Information Retrieval field that chops off the end of words and often includes the removal of derivational affixes. This technique uses the morphology of the language coded in a set of rules that are applied to find out word stems and reduce the vocabulary collapsing derivationally related words. In our study, we use the Snowball Stemmer for Spanish and Italian, and the Porter Stemmer for English that are implemented in NLTK package BIBREF6 . Negation markers might change the polarity of the message. Thus, we attached the negation clue to the nearest word, similar to the approaches used in BIBREF9 . A set of rules was designed for common negation structures that involve negation markers for Spanish, English and Italian. For instance, negation markers used for Spanish are no (not), nunca, jamás (never), and sin (without). The rules (regular expressions) are processed in order, and their purpose is to negate the nearest word to the negation marker using only the information on the text, e.g., avoiding mainly pronouns and articles. For example, in the sentence El coche no es bonito (The car is not nice), the negation marker no and not (for English) is attached to its adjective no_bonito (not_nice). ### Text Representation
After text-transformations, it is needed to represent the text in suitable form in order to use a traditional classifier such as SVM. It was decided to select the well known vector representation of a text given its simplicity and powerful representation. Particularly, it is used the Term Frequency-Inverse Document Frequency which is a well-known weighting scheme in NLP. TF-IDF computes a weight that represents the importance of words or terms inside a document in a collection of documents, i.e., how frequently they appear across multiple documents. Therefore, common words such as the and in, which appear in many documents, will have a low score, and words that appear frequently in a single document will have high score. This weighting scheme selects the terms that represent a document. ### Parameter Optimization
The model selection, sometimes called hyper-parameter optimization, is essential to ensure the performance of a sentiment classifier. In particular, our approach is highly parametric; in fact, we use such property to adapt to several languages. Table TABREF2 summarizes the parameters and their valid values. The search space contains more than 331 thousand configurations when limited to multilingual and language independent parameters; while the search space reaches close to 4 million configurations when we add our three language-dependent parameters. Depending on the size of the training set, each configuration needs several minutes on a commodity server to be evaluated; thus, an exhaustive exploration of the parameter space can be quite expensive making the approach useless in practice. To tackle the efficiency problems, we perform the model selection using two hyper-parameter optimization algorithms. The first corresponds to Random Search, described in depth in BIBREF14 . Random search consists on randomly sampling the parameter space and select the best configuration among the sample. The second algorithm consists on a Hill Climbing BIBREF15 , BIBREF16 implemented with a memory to avoid testing a configuration twice. The main idea behind hill climbing H+M is to take a pivoting configuration, explore the configuration's neighborhood, and greedily moves to the best neighbor. The process is repeated until no improvement is possible. The configuration neighborhood is defined as the set of configurations such that these differ in just one parameter's value. This rule is strengthened for tokenizer (see Table TABREF2 ) to differ in a single internal value not in the whole parameter value. More precisely, let INLINEFORM0 be a valid value for tokenizer and INLINEFORM1 the set of valid values for neighborhoods of INLINEFORM2 , then INLINEFORM3 and INLINEFORM4 for any INLINEFORM5 . To guarantee a better or equal performance than random search, the H+M process starts with the best configuration found in the random search. By using H+M, sample size can be set to 32 or 64, as rule of thumb, and even reach improvements in most cases (see § SECREF4 ). Nonetheless, this simplification and performance boosting comes along with possible higher optimization times. Finally, the performance of each configuration is obtained using a cross-validation technique on the training data, and the metrics are usually used in classification such as: accuracy, score INLINEFORM0 , and recall, among others. ### Datasets and contests
Nowadays, there are several international competitions related to text mining, which include diverse tasks such as: polarity classification (at different levels), subjectivity classification, entity detection, and iron detection, among others. These competitions are relevant to measure the potential of different proposed techniques. In this case, we focused on polarity classification task, hence, we developed a baseline method with an acceptable performance achieved in three different contests, namely, TASS'15 (Spanish) BIBREF17 , SemEval'15-16 (English) BIBREF18 , BIBREF19 , and SENTIPOLC'14 (Italian) BIBREF20 . In addition, our approach was tested with other languages (Arabic, German, Portuguese, Russian, and Swedish) to show that is feasible to use our framework as basis for building more complex sentiment analysis systems. From these languages, datasets and results can be seen in BIBREF21 , BIBREF3 and BIBREF2 . Table TABREF15 presents the details of each of the competitions considered as well as the other languages tested. It can be observed, from the table, the number of examples as well as the number of instances for each polarity level, namely, positive, neutral, negative and none. The training and development (only in SemEval) sets are used to train the sentiment classifier, and the gold set is used to test the classifier. In the case there dataset was not split in training and gold (Arabic, German, Portuguese, Russian, and Swedish) then a cross-validation (10 folds) technique is used to test the classifier. The performance of the classifier is presented using different metrics depending the competition. SemEval uses the average of score INLINEFORM0 of positive and negative labels, TASS uses the accuracy and SENTIPOLC uses a custom metric (see BIBREF17 , BIBREF18 , BIBREF19 , BIBREF20 ). ### Experimental Results
We tested our framework on two kinds of datasets. On one hand, we compare our performance on three languages having well known sentiment analysis contests; here, we compare our work against competitors of those challenges. On the other hand, we selected five languages without popular opinion mining contests; for these languages, we compare our approach against research works reporting the used corpus. ### Performance on sentiment analysis contests
Figure FIGREF17 shows the performance on four contests, corresponding to three different languages. The performance corresponds to the multilingual set of features, i.e., we do not used language-dependent techniques. Figures UID18 - UID21 illustrates the results on each challenge, all competitors are ordered in score's descending order (higher is better). The achieved performance of our approach is marked with a horizontal line on each figure. Figure UID22 briefly describes each challenge and summarizes our performance on each contest; also, we added three standard measures to simplify the insight's creation of the reader. The winner method in SENTIPOLC'14 (Italian) is reported in BIBREF22 . This method uses three groups of features: keyword and micro-blogging characteristics, Sentiment Lexicons, SentiWordNet and MultiWordNet, and Distributional Semantic Model (DSM) with a SVM classifier. In contrast with our method, in BIBREF22 three external sentiment lexicons dictionaries were employed; that is, external information. In TASS'15 (Spanish) competition, the winner reported method was BIBREF23 , which proposed an adaptation based on a tokenizer of tweets Tweetmotif BIBREF24 , Freeling BIBREF25 as lemmatizer, entity detector, morphosyntactic labeler and a translation of the Afinn dictionary. In contrast with our method, BIBREF23 employs several complex and expensive tools. In this task we reached the fourteenth position with an accuracy of INLINEFORM0 . Figure UID19 shows the B4MSA performance to be over two thirds of the competitors. The remaining two contests correspond to the SemEval'15-16. The B4MSA performance in SemEval is depicted in Figures UID20 and UID21 ; here, B4MSA does not perform as well as in other challenges, mainly because, contrary to other challenges, SemEval rules promotes the enrichment of the official training set. To be consistent with the rest of the experiments, B4MSA uses only the official training set. The results can be significantly improved using larger training datasets; for example, joining SemEval'13 and SemEval'16 training sets, we can reach INLINEFORM0 for SemEval'16, which improves the B4MSA's performance (see Table FIGREF17 ). In SemEval'15, the winner method is BIBREF26 , which combines three approaches among the participants of SemEval'13, teams: NRC-Canada, GU-MLT-LT and KLUE, and from SemEval'14 the participant TeamX all of them employing external information. In SemEval'16, the winner method was BIBREF27 is composed with an ensemble of two subsystems based on convolutional neural networks, the first subsystem is created using 290 million tweets, and the second one is feeded with 150 million tweets. All these tweets were selected from a very large unlabeled dataset through distant supervision techniques. Table TABREF23 shows the multilingual set of techniques and the set with language-dependent techniques; for each, we optimized the set of parameters through Random Search and INLINEFORM0 (see Subsection SECREF14 ). The reached performance is reported using both cross-validation and the official gold-standard. Please notice how INLINEFORM1 consistently reaches better performances, even on small sampling sizes. The sampling size is indicated with subscripts in Table TABREF23 . Note that, in SemEval challenges, the cross-validation performances are higher than those reached by evaluating the gold-standard, mainly because the gold-standard does not follow the distribution of training set. This can be understood because the rules of SemEval promote the use of external knowledge. Table TABREF24 compares our performance on five different languages; we do not apply language-dependent techniques. For each comparison, we took a labeled corpus from BIBREF3 (Arabic) and BIBREF21 (the remaining languages). According to author's reports, all tweets were manually labeled by native speakers as pos, neg, or neu. The Arabic dataset contains INLINEFORM0 items; the other datasets contain from 58 thousand tweets to more than 157 thousand tweets. We were able to fetch a fraction of the original datasets; so, we drop the necessary items to hold the original class-population ratio. The ratio of tweets in our training dataset, respect to the original dataset, is indicated beside the name. As before, we evaluate our algorithms through a 10-fold cross validation. In BIBREF3 , BIBREF2 , the authors study the effect of translation in sentiment classifiers; they found better to use native Arabic speakers as annotators than fine-tuned translators plus fine-tuned English sentiment classifiers. In BIBREF21 , the idea is to measure the effect of the agreement among annotators on the production of a sentiment-analysis corpus. Both, on the technical side, both papers use fine tuned classifiers plus a variety of pre-processing techniques to prove their claims. Table TABREF24 supports the idea of choosing B4MSA as a bootstrapping sentiment classifier because, in the overall, B4MSA reaches superior performances regardless of the language. Our approach achieves those performance's levels since it optimizes a set of parameters carefully selected to work on a variety of languages and being robust to informal writing. The latter problem is not properly tackled in many cases. ### Conclusions
We presented a simple to implement multilingual framework for polarity classification whose main contributions are in two aspects. On one hand, our approach can serve as a baseline to compare other classification systems. It considers techniques for text representation such as spelling features, emoticons, word-based n-grams, character-based q-grams and language dependent features. On the other hand, our approach is a framework for practitioners or researchers looking for a bootstrapping sentiment classifier method in order to build more elaborated systems. Besides the text-transformations, the proposed framework uses a SVM classifier (with linear kernel), and, hyper-parameter optimization using random search and H+M over the space of text-transformations. The experimental results show good overall performance in all international contests considered, and the best results in the other five languages tested. It is important to note that all the methods that outperformed B4MSA in the sentiment analysis contests use extra knowledge (lexicons included) meanwhile B4MSA uses only the information provided by each contests. In future work, we will extend our methodology to include extra-knowledge in order to improve the performance. ### Acknowledgements
We would like to thank Valerio Basile, Julio Villena-Roman, and Preslav Nakov for kindly give us access to the gold-standards of SENTIPOLC'14, TASS'15 and SemEval 2015 & 2016, respectively. The authors also thank Elio Villaseñor for the helpful discussions in early stages of this research. Table 1: Parameter list and a brief description of the functionality Table 3: Datasets details from each competition tested in this work Figure 1: The performance listing in four difference challenges. The horizontal lines appearing in a) to d) correspond to B4MSA’s performance. All scores were computed using the official gold-standard and the proper score for each challenge. Table 4: B4MSA’s performance on cross-validation and gold standard. The subscript at right of each score means for the random-search’s parameter (sampling size) needed to find that value. Table 5: Performance on multilingual sentiment analysis (not challenges). B4MSA was restricted to use only the multilingual set of parameters. | text-transformations to the messages, vector space model, Support Vector Machine |
Why was Humphrey being observed?
A. he didn't act like he was expected to
B. to make sure he wasn't a danger to society
C. he was to be observed before he was allowed to be married
D. he was suspected of committing crimes
| A FALL OF GLASS By STANLEY R. LEE Illustrated by DILLON [Transcriber's Note: This etext was produced from Galaxy Magazine October 1960. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed.] The weatherman was always right: Temperature, 59; humidity, 47%; occasional light showers—but of what? The pockets of Mr. Humphrey Fownes were being picked outrageously. It was a splendid day. The temperature was a crisp 59 degrees, the humidity a mildly dessicated 47%. The sun was a flaming orange ball in a cloudless blue sky. His pockets were picked eleven times. It should have been difficult. Under the circumstances it was a masterpiece of pocket picking. What made it possible was Humphrey Fownes' abstraction; he was an uncommonly preoccupied individual. He was strolling along a quiet residential avenue: small private houses, one after another, a place of little traffic and minimum distractions. But he was thinking about weather, which was an unusual subject to begin with for a person living in a domed city. He was thinking so deeply about it that it never occurred to him that entirely too many people were bumping into him. He was thinking about Optimum Dome Conditions (a crisp 59 degrees, a mildly dessicated 47%) when a bogus postman, who pretended to be reading a postal card, jostled him. In the confusion of spilled letters and apologies from both sides, the postman rifled Fownes's handkerchief and inside jacket pockets. He was still thinking about temperature and humidity when a pretty girl happened along with something in her eye. They collided. She got his right and left jacket pockets. It was much too much for coincidence. The sidewalk was wide enough to allow four people to pass at one time. He should surely have become suspicious when two men engaged in a heated argument came along. In the ensuing contretemps they emptied his rear pants pockets, got his wristwatch and restored the contents of the handkerchief pocket. It all went off very smoothly, like a game of put and take—the sole difference being that Humphrey Fownes had no idea he was playing. There was an occasional tinkle of falling glass. It fell on the streets and houses, making small geysers of shiny mist, hitting with a gentle musical sound, like the ephemeral droppings of a celesta. It was precipitation peculiar to a dome: feather-light fragments showering harmlessly on the city from time to time. Dome weevils, their metal arms reaching out with molten glass, roamed the huge casserole, ceaselessly patching and repairing. Humphrey Fownes strode through the puffs of falling glass still intrigued by a temperature that was always 59 degrees, by a humidity that was always 47%, by weather that was always Optimum. It was this rather than skill that enabled the police to maintain such a tight surveillance on him, a surveillance that went to the extent of getting his fingerprints off the postman's bag, and which photographed, X-rayed and chemically analyzed the contents of his pockets before returning them. Two blocks away from his home a careless housewife spilled a five-pound bag of flour as he was passing. It was really plaster of Paris. He left his shoe prints, stride measurement, height, weight and handedness behind. By the time Fownes reached his front door an entire dossier complete with photographs had been prepared and was being read by two men in an orange patrol car parked down the street. Lanfierre had undoubtedly been affected by his job. Sitting behind the wheel of the orange car, he watched Humphrey Fownes approach with a distinct feeling of admiration, although it was an odd, objective kind of admiration, clinical in nature. It was similar to that of a pathologist observing for the first time a new and particularly virulent strain of pneumococcus under his microscope. Lanfierre's job was to ferret out aberration. It couldn't be tolerated within the confines of a dome. Conformity had become more than a social force; it was a physical necessity. And, after years of working at it, Lanfierre had become an admirer of eccentricity. He came to see that genuine quirks were rare and, as time went on, due partly to his own small efforts, rarer. Fownes was a masterpiece of queerness. He was utterly inexplicable. Lanfierre was almost proud of Humphrey Fownes. "Sometimes his house shakes ," Lanfierre said. "House shakes," Lieutenant MacBride wrote in his notebook. Then he stopped and frowned. He reread what he'd just written. "You heard right. The house shakes ," Lanfierre said, savoring it. MacBride looked at the Fownes house through the magnifying glass of the windshield. "Like from ... side to side ?" he asked in a somewhat patronizing tone of voice. "And up and down." MacBride returned the notebook to the breast pocket of his orange uniform. "Go on," he said, amused. "It sounds interesting." He tossed the dossier carelessly on the back seat. Lanfierre sat stiffly behind the wheel, affronted. The cynical MacBride couldn't really appreciate fine aberrations. In some ways MacBride was a barbarian. Lanfierre had held out on Fownes for months. He had even contrived to engage him in conversation once, a pleasantly absurd, irrational little chat that titillated him for weeks. It was only with the greatest reluctance that he finally mentioned Fownes to MacBride. After years of searching for differences Lanfierre had seen how extraordinarily repetitious people were, echoes really, dimly resounding echoes, each believing itself whole and separate. They spoke in an incessant chatter of cliches, and their actions were unbelievably trite. Then a fine robust freak came along and the others—the echoes—refused to believe it. The lieutenant was probably on the point of suggesting a vacation. "Why don't you take a vacation?" Lieutenant MacBride suggested. "It's like this, MacBride. Do you know what a wind is? A breeze? A zephyr?" "I've heard some." "They say there are mountain-tops where winds blow all the time. Strong winds, MacBride. Winds like you and I can't imagine. And if there was a house sitting on such a mountain and if winds did blow, it would shake exactly the way that one does. Sometimes I get the feeling the whole place is going to slide off its foundation and go sailing down the avenue." Lieutenant MacBride pursed his lips. "I'll tell you something else," Lanfierre went on. "The windows all close at the same time. You'll be watching and all of a sudden every single window in the place will drop to its sill." Lanfierre leaned back in the seat, his eyes still on the house. "Sometimes I think there's a whole crowd of people in there waiting for a signal—as if they all had something important to say but had to close the windows first so no one could hear. Why else close the windows in a domed city? And then as soon as the place is buttoned up they all explode into conversation—and that's why the house shakes." MacBride whistled. "No, I don't need a vacation." A falling piece of glass dissolved into a puff of gossamer against the windshield. Lanfierre started and bumped his knee on the steering wheel. "No, you don't need a rest," MacBride said. "You're starting to see flying houses, hear loud babbling voices. You've got winds in your brain, Lanfierre, breezes of fatigue, zephyrs of irrationality—" At that moment, all at once, every last window in the house slammed shut. The street was deserted and quiet, not a movement, not a sound. MacBride and Lanfierre both leaned forward, as if waiting for the ghostly babble of voices to commence. The house began to shake. It rocked from side to side, it pitched forward and back, it yawed and dipped and twisted, straining at the mooring of its foundation. The house could have been preparing to take off and sail down the.... MacBride looked at Lanfierre and Lanfierre looked at MacBride and then they both looked back at the dancing house. "And the water ," Lanfierre said. "The water he uses! He could be the thirstiest and cleanest man in the city. He could have a whole family of thirsty and clean kids, and he still wouldn't need all that water." The lieutenant had picked up the dossier. He thumbed through the pages now in amazement. "Where do you get a guy like this?" he asked. "Did you see what he carries in his pockets?" "And compasses won't work on this street." The lieutenant lit a cigarette and sighed. He usually sighed when making the decision to raid a dwelling. It expressed his weariness and distaste for people who went off and got neurotic when they could be enjoying a happy, normal existence. There was something implacable about his sighs. "He'll be coming out soon," Lanfierre said. "He eats supper next door with a widow. Then he goes to the library. Always the same. Supper at the widow's next door and then the library." MacBride's eyebrows went up a fraction of an inch. "The library?" he said. "Is he in with that bunch?" Lanfierre nodded. "Should be very interesting," MacBride said slowly. "I can't wait to see what he's got in there," Lanfierre murmured, watching the house with a consuming interest. They sat there smoking in silence and every now and then their eyes widened as the house danced a new step. Fownes stopped on the porch to brush the plaster of paris off his shoes. He hadn't seen the patrol car and this intense preoccupation of his was also responsible for the dancing house—he simply hadn't noticed. There was a certain amount of vibration, of course. He had a bootleg pipe connected into the dome blower system, and the high-pressure air caused some buffeting against the thin walls of the house. At least, he called it buffeting; he'd never thought to watch from outside. He went in and threw his jacket on the sofa, there being no room left in the closets. Crossing the living room he stopped to twist a draw-pull. Every window slammed shut. "Tight as a kite," he thought, satisfied. He continued on toward the closet at the foot of the stairs and then stopped again. Was that right? No, snug as a hug in a rug . He went on, thinking: The old devils. The downstairs closet was like a great watch case, a profusion of wheels surrounding the Master Mechanism, which was a miniature see-saw that went back and forth 365-1/4 times an hour. The wheels had a curious stateliness about them. They were all quite old, salvaged from grandfather's clocks and music boxes and they went around in graceful circles at the rate of 30 and 31 times an hour ... although there was one slightly eccentric cam that vacillated between 28 and 29. He watched as they spun and flashed in the darkness, and then set them for seven o'clock in the evening, April seventh, any year. Outside, the domed city vanished. It was replaced by an illusion. Or, as Fownes hoped it might appear, the illusion of the domed city vanished and was replaced by a more satisfactory, and, for his specific purpose, more functional, illusion. Looking through the window he saw only a garden. Instead of an orange sun at perpetual high noon, there was a red sun setting brilliantly, marred only by an occasional arcover which left the smell of ozone in the air. There was also a gigantic moon. It hid a huge area of sky, and it sang. The sun and moon both looked down upon a garden that was itself scintillant, composed largely of neon roses. Moonlight, he thought, and roses. Satisfactory. And cocktails for two. Blast, he'd never be able to figure that one out! He watched as the moon played, Oh, You Beautiful Doll and the neon roses flashed slowly from red to violet, then went back to the closet and turned on the scent. The house began to smell like an immensely concentrated rose as the moon shifted to People Will Say We're In Love . He rubbed his chin critically. It seemed all right. A dreamy sunset, an enchanted moon, flowers, scent. They were all purely speculative of course. He had no idea how a rose really smelled—or looked for that matter. Not to mention a moon. But then, neither did the widow. He'd have to be confident, assertive. Insist on it. I tell you, my dear, this is a genuine realistic romantic moon. Now, does it do anything to your pulse? Do you feel icy fingers marching up and down your spine? His own spine didn't seem to be affected. But then he hadn't read that book on ancient mores and courtship customs. How really odd the ancients were. Seduction seemed to be an incredibly long and drawn-out process, accompanied by a considerable amount of falsification. Communication seemed virtually impossible. "No" meant any number of things, depending on the tone of voice and the circumstances. It could mean yes, it could mean ask me again later on this evening. He went up the stairs to the bedroom closet and tried the rain-maker, thinking roguishly: Thou shalt not inundate. The risks he was taking! A shower fell gently on the garden and a male chorus began to chant Singing in the Rain . Undiminished, the yellow moon and the red sun continued to be brilliant, although the sun occasionally arced over and demolished several of the neon roses. The last wheel in the bedroom closet was a rather elegant steering wheel from an old 1995 Studebaker. This was on the bootleg pipe; he gingerly turned it. Far below in the cellar there was a rumble and then the soft whistle of winds came to him. He went downstairs to watch out the living room window. This was important; the window had a really fixed attitude about air currents. The neon roses bent and tinkled against each other as the wind rose and the moon shook a trifle as it whispered Cuddle Up a Little Closer . He watched with folded arms, considering how he would start. My dear Mrs. Deshazaway. Too formal. They'd be looking out at the romantic garden; time to be a bit forward. My very dear Mrs. Deshazaway. No. Contrived. How about a simple, Dear Mrs. Deshazaway . That might be it. I was wondering, seeing as how it's so late, if you wouldn't rather stay over instead of going home.... Preoccupied, he hadn't noticed the winds building up, didn't hear the shaking and rattling of the pipes. There were attic pipes connected to wall pipes and wall pipes connected to cellar pipes, and they made one gigantic skeleton that began to rattle its bones and dance as high-pressure air from the dome blower rushed in, slowly opening the Studebaker valve wider and wider.... The neon roses thrashed about, extinguishing each other. The red sun shot off a mass of sparks and then quickly sank out of sight. The moon fell on the garden and rolled ponderously along, crooning When the Blue of the Night Meets the Gold of the Day . The shaking house finally woke him up. He scrambled upstairs to the Studebaker wheel and shut it off. At the window again, he sighed. Repairs were in order. And it wasn't the first time the winds got out of line. Why didn't she marry him and save all this bother? He shut it all down and went out the front door, wondering about the rhyme of the months, about stately August and eccentric February and romantic April. April. Its days were thirty and it followed September. And all the rest have thirty-one. What a strange people, the ancients! He still didn't see the orange car parked down the street. "Men are too perishable," Mrs. Deshazaway said over dinner. "For all practical purposes I'm never going to marry again. All my husbands die." "Would you pass the beets, please?" Humphrey Fownes said. She handed him a platter of steaming red beets. "And don't look at me that way," she said. "I'm not going to marry you and if you want reasons I'll give you four of them. Andrew. Curt. Norman. And Alphonse." The widow was a passionate woman. She did everything passionately—talking, cooking, dressing. Her beets were passionately red. Her clothes rustled and her high heels clicked and her jewelry tinkled. She was possessed by an uncontrollable dynamism. Fownes had never known anyone like her. "You forgot to put salt on the potatoes," she said passionately, then went on as calmly as it was possible for her to be, to explain why she couldn't marry him. "Do you have any idea what people are saying? They're all saying I'm a cannibal! I rob my husbands of their life force and when they're empty I carry their bodies outside on my way to the justice of the peace." "As long as there are people," he said philosophically, "there'll be talk." "But it's the air! Why don't they talk about that? The air is stale, I'm positive. It's not nourishing. The air is stale and Andrew, Curt, Norman and Alphonse couldn't stand it. Poor Alphonse. He was never so healthy as on the day he was born. From then on things got steadily worse for him." "I don't seem to mind the air." She threw up her hands. "You'd be the worst of the lot!" She left the table, rustling and tinkling about the room. "I can just hear them. Try some of the asparagus. Five. That's what they'd say. That woman did it again. And the plain fact is I don't want you on my record." "Really," Fownes protested. "I feel splendid. Never better." He could hear her moving about and then felt her hands on his shoulders. "And what about those very elaborate plans you've been making to seduce me?" Fownes froze with three asparagus hanging from his fork. "Don't you think they'll find out? I found out and you can bet they will. It's my fault, I guess. I talk too much. And I don't always tell the truth. To be completely honest with you, Mr. Fownes, it wasn't the old customs at all standing between us, it was air. I can't have another man die on me, it's bad for my self-esteem. And now you've gone and done something good and criminal, something peculiar." Fownes put his fork down. "Dear Mrs. Deshazaway," he started to say. "And of course when they do find out and they ask you why, Mr. Fownes, you'll tell them. No, no heroics, please! When they ask a man a question he always answers and you will too. You'll tell them I wanted to be courted and when they hear that they'll be around to ask me a few questions. You see, we're both a bit queer." "I hadn't thought of that," Fownes said quietly. "Oh, it doesn't really matter. I'll join Andrew, Curt, Norman—" "That won't be necessary," Fownes said with unusual force. "With all due respect to Andrew, Curt, Norman and Alphonse, I might as well state here and now I have other plans for you, Mrs. Deshazaway." "But my dear Mr. Fownes," she said, leaning across the table. "We're lost, you and I." "Not if we could leave the dome," Fownes said quietly. "That's impossible! How?" In no hurry, now that he had the widow's complete attention, Fownes leaned across the table and whispered: "Fresh air, Mrs. Deshazaway? Space? Miles and miles of space where the real-estate monopoly has no control whatever? Where the wind blows across prairies ; or is it the other way around? No matter. How would you like that , Mrs. Deshazaway?" Breathing somewhat faster than usual, the widow rested her chin on her two hands. "Pray continue," she said. "Endless vistas of moonlight and roses? April showers, Mrs. Deshazaway. And June, which as you may know follows directly upon April and is supposed to be the month of brides, of marrying. June also lies beyond the dome." "I see." " And ," Mr. Fownes added, his voice a honeyed whisper, "they say that somewhere out in the space and the roses and the moonlight, the sleeping equinox yawns and rises because on a certain day it's vernal and that's when it roams the Open Country where geigers no longer scintillate." " My. " Mrs. Deshazaway rose, paced slowly to the window and then came back to the table, standing directly over Fownes. "If you can get us outside the dome," she said, "out where a man stays warm long enough for his wife to get to know him ... if you can do that, Mr. Fownes ... you may call me Agnes." When Humphrey Fownes stepped out of the widow's house, there was a look of such intense abstraction on his features that Lanfierre felt a wistful desire to get out of the car and walk along with the man. It would be such a deliciously insane experience. ("April has thirty days," Fownes mumbled, passing them, "because thirty is the largest number such that all smaller numbers not having a common divisor with it are primes ." MacBride frowned and added it to the dossier. Lanfierre sighed.) Pinning his hopes on the Movement, Fownes went straight to the library several blocks away, a shattered depressing place given over to government publications and censored old books with holes in them. It was used so infrequently that the Movement was able to meet there undisturbed. The librarian was a yellowed, dog-eared woman of eighty. She spent her days reading ancient library cards and, like the books around her, had been rendered by time's own censor into near unintelligibility. "Here's one," she said to him as he entered. " Gulliver's Travels. Loaned to John Wesley Davidson on March 14, 1979 for five days. What do you make of it?" In the litter of books and cards and dried out ink pads that surrounded the librarian, Fownes noticed a torn dust jacket with a curious illustration. "What's that?" he said. "A twister," she replied quickly. "Now listen to this . Seven years later on March 21, 1986, Ella Marshall Davidson took out the same book. What do you make of that ?" "I'd say," Humphrey Fownes said, "that he ... that he recommended it to her, that one day they met in the street and he told her about this book and then they ... they went to the library together and she borrowed it and eventually, why eventually they got married." "Hah! They were brother and sister!" the librarian shouted in her parched voice, her old buckram eyes laughing with cunning. Fownes smiled weakly and looked again at the dust jacket. The twister was unquestionably a meteorological phenomenon. It spun ominously, like a malevolent top, and coursed the countryside destructively, carrying a Dorothy to an Oz. He couldn't help wondering if twisters did anything to feminine pulses, if they could possibly be a part of a moonlit night, with cocktails and roses. He absently stuffed the dust jacket in his pocket and went on into the other rooms, the librarian mumbling after him: "Edna Murdoch Featherstone, April 21, 1991," as though reading inscriptions on a tombstone. The Movement met in what had been the children's room, where unpaid ladies of the afternoon had once upon a time read stories to other people's offspring. The members sat around at the miniature tables looking oddly like giants fled from their fairy tales, protesting. "Where did the old society fail?" the leader was demanding of them. He stood in the center of the room, leaning on a heavy knobbed cane. He glanced around at the group almost complacently, and waited as Humphrey Fownes squeezed into an empty chair. "We live in a dome," the leader said, "for lack of something. An invention! What is the one thing that the great technological societies before ours could not invent, notwithstanding their various giant brains, electronic and otherwise?" Fownes was the kind of man who never answered a rhetorical question. He waited, uncomfortable in the tight chair, while the others struggled with this problem in revolutionary dialectics. " A sound foreign policy ," the leader said, aware that no one else had obtained the insight. "If a sound foreign policy can't be created the only alternative is not to have any foreign policy at all. Thus the movement into domes began— by common consent of the governments . This is known as self-containment." Dialectically out in left field, Humphrey Fownes waited for a lull in the ensuing discussion and then politely inquired how it might be arranged for him to get out. "Out?" the leader said, frowning. "Out? Out where?" "Outside the dome." "Oh. All in good time, my friend. One day we shall all pick up and leave." "And that day I'll await impatiently," Fownes replied with marvelous tact, "because it will be lonely out there for the two of us. My future wife and I have to leave now ." "Nonsense. Ridiculous! You have to be prepared for the Open Country. You can't just up and leave, it would be suicide, Fownes. And dialectically very poor." "Then you have discussed preparations, the practical necessities of life in the Open Country. Food, clothing, a weapon perhaps? What else? Have I left anything out?" The leader sighed. "The gentleman wants to know if he's left anything out," he said to the group. Fownes looked around at them, at some dozen pained expressions. "Tell the man what he's forgotten," the leader said, walking to the far window and turning his back quite pointedly on them. Everyone spoke at the same moment. " A sound foreign policy ," they all said, it being almost too obvious for words. On his way out the librarian shouted at him: " A Tale of a Tub , thirty-five years overdue!" She was calculating the fine as he closed the door. Humphrey Fownes' preoccupation finally came to an end when he was one block away from his house. It was then that he realized something unusual must have occurred. An orange patrol car of the security police was parked at his front door. And something else was happening too. His house was dancing. It was disconcerting, and at the same time enchanting, to watch one's residence frisking about on its foundation. It was such a strange sight that for the moment he didn't give a thought to what might be causing it. But when he stepped gingerly onto the porch, which was doing its own independent gavotte, he reached for the doorknob with an immense curiosity. The door flung itself open and knocked him back off the porch. From a prone position on his miniscule front lawn, Fownes watched as his favorite easy chair sailed out of the living room on a blast of cold air and went pinwheeling down the avenue in the bright sunshine. A wild wind and a thick fog poured out of the house. It brought chairs, suits, small tables, lamps trailing their cords, ashtrays, sofa cushions. The house was emptying itself fiercely, as if disgorging an old, spoiled meal. From deep inside he could hear the rumble of his ancient upright piano as it rolled ponderously from room to room. He stood up; a wet wind swept over him, whipping at his face, toying with his hair. It was a whistling in his ears, and a tingle on his cheeks. He got hit by a shoe. As he forced his way back to the doorway needles of rain played over his face and he heard a voice cry out from somewhere in the living room. "Help!" Lieutenant MacBride called. Standing in the doorway with his wet hair plastered down on his dripping scalp, the wind roaring about him, the piano rumbling in the distance like thunder, Humphrey Fownes suddenly saw it all very clearly. " Winds ," he said in a whisper. "What's happening?" MacBride yelled, crouching behind the sofa. " March winds," he said. "What?!" "April showers!" The winds roared for a moment and then MacBride's lost voice emerged from the blackness of the living room. "These are not Optimum Dome Conditions!" the voice wailed. "The temperature is not 59 degrees. The humidity is not 47%!" Fownes held his face up to let the rain fall on it. "Moonlight!" he shouted. "Roses! My soul for a cocktail for two!" He grasped the doorway to keep from being blown out of the house. "Are you going to make it stop or aren't you!" MacBride yelled. "You'll have to tell me what you did first!" "I told him not to touch that wheel! Lanfierre. He's in the upstairs bedroom!" When he heard this Fownes plunged into the house and fought his way up the stairs. He found Lanfierre standing outside the bedroom with a wheel in his hand. "What have I done?" Lanfierre asked in the monotone of shock. Fownes took the wheel. It was off a 1995 Studebaker. "I'm not sure what's going to come of this," he said to Lanfierre with an astonishing amount of objectivity, "but the entire dome air supply is now coming through my bedroom." The wind screamed. "Is there something I can turn?" Lanfierre asked. "Not any more there isn't." They started down the stairs carefully, but the wind caught them and they quickly reached the bottom in a wet heap. Recruiting Lieutenant MacBride from behind his sofa, the men carefully edged out of the house and forced the front door shut. The wind died. The fog dispersed. They stood dripping in the Optimum Dome Conditions of the bright avenue. "I never figured on this ," Lanfierre said, shaking his head. With the front door closed the wind quickly built up inside the house. They could see the furnishing whirl past the windows. The house did a wild, elated jig. "What kind of a place is this?" MacBride said, his courage beginning to return. He took out his notebook but it was a soggy mess. He tossed it away. "Sure, he was different ," Lanfierre murmured. "I knew that much." When the roof blew off they weren't really surprised. With a certain amount of equanimity they watched it lift off almost gracefully, standing on end for a moment before toppling to the ground. It was strangely slow motion, as was the black twirling cloud that now rose out of the master bedroom, spewing shorts and socks and cases every which way. " Now what?" MacBride said, thoroughly exasperated, as this strange black cloud began to accelerate, whirling about like some malevolent top.... Humphrey Fownes took out the dust jacket he'd found in the library. He held it up and carefully compared the spinning cloud in his bedroom with the illustration. The cloud rose and spun, assuming the identical shape of the illustration. "It's a twister," he said softly. "A Kansas twister!" "What," MacBride asked, his bravado slipping away again, "what ... is a twister?" The twister roared and moved out of the bedroom, out over the rear of the house toward the side of the dome. "It says here," Fownes shouted over the roaring, "that Dorothy traveled from Kansas to Oz in a twister and that ... and that Oz is a wonderful and mysterious land beyond the confines of everyday living ." MacBride's eyes and mouth were great zeros. "Is there something I can turn?" Lanfierre asked. Huge chunks of glass began to fall around them. "Fownes!" MacBride shouted. "This is a direct order! Make it go back!" But Fownes had already begun to run on toward the next house, dodging mountainous puffs of glass as he went. "Mrs. Deshazaway!" he shouted. "Yoo-hoo, Mrs. Deshazaway!" The dome weevils were going berserk trying to keep up with the precipitation. They whirred back and forth at frightful speed, then, emptied of molten glass, rushed to the Trough which they quickly emptied and then rushed about empty-handed. "Yoo-hoo!" he yelled, running. The artificial sun vanished behind the mushrooming twister. Optimum temperature collapsed. "Mrs. Deshazaway! Agnes , will you marry me? Yoo-hoo!" Lanfierre and Lieutenant MacBride leaned against their car and waited, dazed. There was quite a large fall of glass. | A. he didn't act like he was expected to |
What don't Mia and Ri have in common?
A. they both think Extrone is going to kill them
B. they've killed farn beasts
C. they're businessmen
D. they both dislike Extrone
| HUNT the HUNTER BY KRIS NEVILLE Illustrated by ELIZABETH MacINTYRE [Transcriber's Note: This etext was produced from Galaxy Science Fiction June 1951. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed.] Of course using live bait is the best way to lure dangerous alien animals ... unless it turns out that you are the bait! "We're somewhat to the south, I think," Ri said, bending over the crude field map. "That ridge," he pointed, "on our left, is right here." He drew a finger down the map. "It was over here," he moved the finger, "over the ridge, north of here, that we sighted them." Extrone asked, "Is there a pass?" Ri looked up, studying the terrain. He moved his shoulders. "I don't know, but maybe they range this far. Maybe they're on this side of the ridge, too." Delicately, Extrone raised a hand to his beard. "I'd hate to lose a day crossing the ridge," he said. "Yes, sir," Ri said. Suddenly he threw back his head. "Listen!" "Eh?" Extrone said. "Hear it? That cough? I think that's one, from over there. Right up ahead of us." Extrone raised his eyebrows. This time, the coughing roar was more distant, but distinct. "It is!" Ri said. "It's a farn beast, all right!" Extrone smiled, almost pointed teeth showing through the beard. "I'm glad we won't have to cross the ridge." Ri wiped his forehead on the back of his sleeve. "Yes, sir." "We'll pitch camp right here, then," Extrone said. "We'll go after it tomorrow." He looked at the sky. "Have the bearers hurry." "Yes, sir." Ri moved away, his pulse gradually slowing. "You, there!" he called. "Pitch camp, here!" He crossed to Mia, who, along with him, had been pressed into Extrone's party as guides. Once more, Ri addressed the bearers, "Be quick, now!" And to Mia, "God almighty, he was getting mad." He ran a hand under his collar. "It's a good thing that farn beast sounded off when it did. I'd hate to think of making him climb that ridge." Mia glanced nervously over his shoulder. "It's that damned pilot's fault for setting us down on this side. I told him it was the other side. I told him so." Ri shrugged hopelessly. Mia said, "I don't think he even saw a blast area over here. I think he wanted to get us in trouble." "There shouldn't be one. There shouldn't be a blast area on this side of the ridge, too." "That's what I mean. The pilot don't like businessmen. He had it in for us." Ri cleared his throat nervously. "Maybe you're right." "It's the Hunting Club he don't like." "I wish to God I'd never heard of a farn beast," Ri said. "At least, then, I wouldn't be one of his guides. Why didn't he hire somebody else?" Mia looked at his companion. He spat. "What hurts most, he pays us for it. I could buy half this planet, and he makes me his guide—at less than I pay my secretary." "Well, anyway, we won't have to cross that ridge." "Hey, you!" Extrone called. The two of them turned immediately. "You two scout ahead," Extrone said. "See if you can pick up some tracks." "Yes, sir," Ri said, and instantly the two of them readjusted their shoulder straps and started off. Shortly they were inside of the scrub forest, safe from sight. "Let's wait here," Mia said. "No, we better go on. He may have sent a spy in." They pushed on, being careful to blaze the trees, because they were not professional guides. "We don't want to get too near," Ri said after toiling through the forest for many minutes. "Without guns, we don't want to get near enough for the farn beast to charge us." They stopped. The forest was dense, the vines clinging. "He'll want the bearers to hack a path for him," Mia said. "But we go it alone. Damn him." Ri twisted his mouth into a sour frown. He wiped at his forehead. "Hot. By God, it's hot. I didn't think it was this hot, the first time we were here." Mia said, "The first time, we weren't guides. We didn't notice it so much then." They fought a few yards more into the forest. Then it ended. Or, rather, there was a wide gap. Before them lay a blast area, unmistakable. The grass was beginning to grow again, but the tree stumps were roasted from the rocket breath. "This isn't ours!" Ri said. "This looks like it was made nearly a year ago!" Mia's eyes narrowed. "The military from Xnile?" "No," Ri said. "They don't have any rockets this small. And I don't think there's another cargo rocket on this planet outside of the one we leased from the Club. Except the one he brought." "The ones who discovered the farn beasts in the first place?" Mia asked. "You think it's their blast?" "So?" Ri said. "But who are they?" It was Mia's turn to shrug. "Whoever they were, they couldn't have been hunters. They'd have kept the secret better." "We didn't do so damned well." "We didn't have a chance," Mia objected. "Everybody and his brother had heard the rumor that farn beasts were somewhere around here. It wasn't our fault Extrone found out." "I wish we hadn't shot our guide, then. I wish he was here instead of us." Mia shook perspiration out of his eyes. "We should have shot our pilot, too. That was our mistake. The pilot must have been the one who told Extrone we'd hunted this area." "I didn't think a Club pilot would do that." "After Extrone said he'd hunt farn beasts, even if it meant going to the alien system? Listen, you don't know.... Wait a minute." There was perspiration on Ri's upper lip. " I didn't tell Extrone, if that's what you're thinking," Mia said. Ri's mouth twisted. "I didn't say you did." "Listen," Mia said in a hoarse whisper. "I just thought. Listen. To hell with how he found out. Here's the point. Maybe he'll shoot us, too, when the hunt's over." Ri licked his lips. "No. He wouldn't do that. We're not—not just anybody. He couldn't kill us like that. Not even him . And besides, why would he want to do that? It wouldn't do any good to shoot us. Too many people already know about the farn beasts. You said that yourself." Mia said, "I hope you're right." They stood side by side, studying the blast area in silence. Finally, Mia said, "We better be getting back." "What'll we tell him?" "That we saw tracks. What else can we tell him?" They turned back along their trail, stumbling over vines. "It gets hotter at sunset," Ri said nervously. "The breeze dies down." "It's screwy. I didn't think farn beasts had this wide a range. There must be a lot of them, to be on both sides of the ridge like this." "There may be a pass," Mia said, pushing a vine away. Ri wrinkled his brow, panting. "I guess that's it. If there were a lot of them, we'd have heard something before we did. But even so, it's damned funny, when you think about it." Mia looked up at the darkening sky. "We better hurry," he said. When it came over the hastily established camp, the rocket was low, obviously looking for a landing site. It was a military craft, from the outpost on the near moon, and forward, near the nose, there was the blazoned emblem of the Ninth Fleet. The rocket roared directly over Extrone's tent, turned slowly, spouting fuel expensively, and settled into the scrub forest, turning the vegetation beneath it sere by its blasts. Extrone sat on an upholstered stool before his tent and spat disgustedly and combed his beard with his blunt fingers. Shortly, from the direction of the rocket, a group of four high-ranking officers came out of the forest, heading toward him. They were spruce, the officers, with military discipline holding their waists in and knees almost stiff. "What in hell do you want?" Extrone asked. They stopped a respectful distance away. "Sir...." one began. "Haven't I told you gentlemen that rockets frighten the game?" Extrone demanded, ominously not raising his voice. "Sir," the lead officer said, "it's another alien ship. It was sighted a few hours ago, off this very planet, sir." Extrone's face looked much too innocent. "How did it get there, gentlemen? Why wasn't it destroyed?" "We lost it again, sir. Temporarily, sir." "So?" Extrone mocked. "We thought you ought to return to a safer planet, sir. Until we could locate and destroy it." Extrone stared at them for a space. Then, indifferently, he turned away, in the direction of a resting bearer. "You!" he said. "Hey! Bring me a drink!" He faced the officers again. He smiled maliciously. "I'm staying here." The lead officer licked his firm lower lip. "But, sir...." Extrone toyed with his beard. "About a year ago, gentlemen, there was an alien ship around here then, wasn't there? And you destroyed it, didn't you?" "Yes, sir. When we located it, sir." "You'll destroy this one, too," Extrone said. "We have a tight patrol, sir. It can't slip through. But it might try a long range bombardment, sir." Extrone said, "To begin with, they probably don't even know I'm here. And they probably couldn't hit this area if they did know. And you can't afford to let them get a shot at me, anyway." "That's why we'd like you to return to an inner planet, sir." Extrone plucked at his right ear lobe, half closing his eyes. "You'll lose a fleet before you'll dare let anything happen to me, gentlemen. I'm quite safe here, I think." The bearer brought Extrone his drink. "Get off," Extrone said quietly to the four officers. Again they turned reluctantly. This time, he did not call them back. Instead, with amusement, he watched until they disappeared into the tangle of forest. Dusk was falling. The takeoff blast of the rocket illuminated the area, casting weird shadows on the gently swaying grasses; there was a hot breath of dry air and the rocket dwindled toward the stars. Extrone stood up lazily, stretching. He tossed the empty glass away, listened for it to shatter. He reached out, parted the heavy flap to his tent. "Sir?" Ri said, hurrying toward him in the gathering darkness. "Eh?" Extrone said, turning, startled. "Oh, you. Well?" "We ... located signs of the farn beast, sir. To the east." Extrone nodded. After a moment he said, "You killed one, I believe, on your trip?" Ri shifted. "Yes, sir." Extrone held back the flap of the tent. "Won't you come in?" he asked without any politeness whatever. Ri obeyed the order. The inside of the tent was luxurious. The bed was of bulky feathers, costly of transport space, the sleep curtains of silken gauze. The floor, heavy, portable tile blocks, not the hollow kind, were neatly and smoothly inset into the ground. Hanging from the center, to the left of the slender, hand-carved center pole, was a chain of crystals. They tinkled lightly when Extrone dropped the flap. The light was electric from a portable dynamo. Extrone flipped it on. He crossed to the bed, sat down. "You were, I believe, the first ever to kill a farn beast?" he said. "I.... No, sir. There must have been previous hunters, sir." Extrone narrowed his eyes. "I see by your eyes that you are envious—that is the word, isn't it?—of my tent." Ri looked away from his face. "Perhaps I'm envious of your reputation as a hunter. You see, I have never killed a farn beast. In fact, I haven't seen a farn beast." Ri glanced nervously around the tent, his sharp eyes avoiding Extrone's glittering ones. "Few people have seen them, sir." "Oh?" Extrone questioned mildly. "I wouldn't say that. I understand that the aliens hunt them quite extensively ... on some of their planets." "I meant in our system, sir." "Of course you did," Extrone said, lazily tracing the crease of his sleeve with his forefinger. "I imagine these are the only farn beasts in our system." Ri waited uneasily, not answering. "Yes," Extrone said, "I imagine they are. It would have been a shame if you had killed the last one. Don't you think so?" Ri's hands worried the sides of his outer garment. "Yes, sir. It would have been." Extrone pursed his lips. "It wouldn't have been very considerate of you to—But, still, you gained valuable experience. I'm glad you agreed to come along as my guide." "It was an honor, sir." Extrone's lip twisted in wry amusement. "If I had waited until it was safe for me to hunt on an alien planet, I would not have been able to find such an illustrious guide." "... I'm flattered, sir." "Of course," Extrone said. "But you should have spoken to me about it, when you discovered the farn beast in our own system." "I realize that, sir. That is, I had intended at the first opportunity, sir...." "Of course," Extrone said dryly. "Like all of my subjects," he waved his hand in a broad gesture, "the highest as well as the lowest slave, know me and love me. I know your intentions were the best." Ri squirmed, his face pale. "We do indeed love you, sir." Extrone bent forward. " Know me and love me." "Yes, sir. Know you and love you, sir," Ri said. "Get out!" Extrone said. "It's frightening," Ri said, "to be that close to him." Mia nodded. The two of them, beneath the leaf-swollen branches of the gnarled tree, were seated on their sleeping bags. The moon was clear and cold and bright in a cloudless sky; a small moon, smooth-surfaced, except for a central mountain ridge that bisected it into almost twin hemispheres. "To think of him. As flesh and blood. Not like the—well; that—what we've read about." Mia glanced suspiciously around him at the shadows. "You begin to understand a lot of things, after seeing him." Ri picked nervously at the cover of his sleeping bag. "It makes you think," Mia added. He twitched. "I'm afraid. I'm afraid he'll.... Listen, we'll talk. When we get back to civilization. You, me, the bearers. About him. He can't let that happen. He'll kill us first." Ri looked up at the moon, shivering. "No. We have friends. We have influence. He couldn't just like that—" "He could say it was an accident." "No," Ri said stubbornly. "He can say anything," Mia insisted. "He can make people believe anything. Whatever he says. There's no way to check on it." "It's getting cold," Ri said. "Listen," Mia pleaded. "No," Ri said. "Even if we tried to tell them, they wouldn't listen. Everybody would know we were lying. Everything they've come to believe would tell them we were lying. Everything they've read, every picture they've seen. They wouldn't believe us. He knows that." "Listen," Mia repeated intently. "This is important. Right now he couldn't afford to let us talk. Not right now. Because the Army is not against him. Some officers were here, just before we came back. A bearer overheard them talking. They don't want to overthrow him!" Ri's teeth, suddenly, were chattering. "That's another lie," Mia continued. "That he protects the people from the Army. That's a lie. I don't believe they were ever plotting against him. Not even at first. I think they helped him, don't you see?" Ri whined nervously. "It's like this," Mia said. "I see it like this. The Army put him in power when the people were in rebellion against military rule." Ri swallowed. "We couldn't make the people believe that." "No?" Mia challenged. "Couldn't we? Not today, but what about tomorrow? You'll see. Because I think the Army is getting ready to invade the alien system!" "The people won't support them," Ri answered woodenly. " Think. If he tells them to, they will. They trust him." Ri looked around at the shadows. "That explains a lot of things," Mia said. "I think the Army's been preparing for this for a long time. From the first, maybe. That's why Extrone cut off our trade with the aliens. Partly to keep them from learning that he was getting ready to invade them, but more to keep them from exposing him to the people. The aliens wouldn't be fooled like we were, so easy." "No!" Ri snapped. "It was to keep the natural economic balance." "You know that's not right." Ri lay down on his bed roll. "Don't talk about it. It's not good to talk like this. I don't even want to listen." "When the invasion starts, he'll have to command all their loyalties. To keep them from revolt again. They'd be ready to believe us, then. He'll have a hard enough time without people running around trying to tell the truth." "You're wrong. He's not like that. I know you're wrong." Mia smiled twistedly. "How many has he already killed? How can we even guess?" Ri swallowed sickly. "Remember our guide? To keep our hunting territory a secret?" Ri shuddered. "That's different. Don't you see? This is not at all like that." With morning came birds' songs, came dew, came breakfast smells. The air was sweet with cooking and it was nostalgic, childhoodlike, uncontaminated. And Extrone stepped out of the tent, fully dressed, surly, letting the flap slap loudly behind him. He stretched hungrily and stared around the camp, his eyes still vacant-mean with sleep. "Breakfast!" he shouted, and two bearers came running with a folding table and chair. Behind them, a third bearer, carrying a tray of various foods; and yet behind him, a fourth, with a steaming pitcher and a drinking mug. Extrone ate hugely, with none of the delicacy sometimes affected in his conversational gestures. When he had finished, he washed his mouth with water and spat on the ground. "Lin!" he said. His personal bearer came loping toward him. "Have you read that manual I gave you?" Lin nodded. "Yes." Extrone pushed the table away. He smacked his lips wetly. "Very ludicrous, Lin. Have you noticed that I have two businessmen for guides? It occurred to me when I got up. They would have spat on me, twenty years ago, damn them." Lin waited. "Now I can spit on them, which pleases me." "The farn beasts are dangerous, sir," Lin said. "Eh? Oh, yes. Those. What did the manual say about them?" "I believe they're carnivorous, sir." "An alien manual. That's ludicrous, too. That we have the only information on our newly discovered fauna from an alien manual—and, of course, two businessmen." "They have very long, sharp fangs, and, when enraged, are capable of tearing a man—" "An alien?" Extrone corrected. "There's not enough difference between us to matter, sir. Of tearing an alien to pieces, sir." Extrone laughed harshly. "It's 'sir' whenever you contradict me?" Lin's face remained impassive. "I guess it seems that way. Sir." "Damned few people would dare go as far as you do," Extrone said. "But you're afraid of me, too, in your own way, aren't you?" Lin shrugged. "Maybe." "I can see you are. Even my wives are. I wonder if anyone can know how wonderful it feels to have people all afraid of you." "The farn beasts, according to the manual...." "You are very insistent on one subject." "... It's the only thing I know anything about. The farn beast, as I was saying, sir, is the particular enemy of men. Or if you like, of aliens. Sir." "All right," Extrone said, annoyed. "I'll be careful." In the distance, a farn beast coughed. Instantly alert, Extrone said, "Get the bearers! Have some of them cut a path through that damn thicket! And tell those two businessmen to get the hell over here!" Lin smiled, his eyes suddenly afire with the excitement of the hunt. Four hours later, they were well into the scrub forest. Extrone walked leisurely, well back of the cutters, who hacked away, methodically, at the vines and branches which might impede his forward progress. Their sharp, awkward knives snickered rhythmically to the rasp of their heavy breathing. Occasionally, Extrone halted, motioned for his water carrier, and drank deeply of the icy water to allay the heat of the forest, a heat made oppressive by the press of foliage against the outside air. Ranging out, on both sides of the central body, the two businessmen fought independently against the wild growth, each scouting the flanks for farn beasts, and ahead, beyond the cutters, Lin flittered among the tree trunks, sometimes far, sometimes near. Extrone carried the only weapon, slung easily over his shoulder, a powerful blast rifle, capable of piercing medium armor in sustained fire. To his rear, the water carrier was trailed by a man bearing a folding stool, and behind him, a man carrying the heavy, high-powered two-way communication set. Once Extrone unslung his blast rifle and triggered a burst at a tiny, arboreal mammal, which, upon the impact, shattered asunder, to Extrone's satisfied chuckle, in a burst of blood and fur. When the sun stood high and heat exhaustion made the near-naked bearers slump, Extrone permitted a rest. While waiting for the march to resume, he sat on the stool with his back against an ancient tree and patted, reflectively, the blast rifle, lying across his legs. "For you, sir," the communications man said, interrupting his reverie. "Damn," Extrone muttered. His face twisted in anger. "It better be important." He took the head-set and mike and nodded to the bearer. The bearer twiddled the dials. "Extrone. Eh?... Oh, you got their ship. Well, why in hell bother me?... All right, so they found out I was here. You got them, didn't you?" "Blasted them right out of space," the voice crackled excitedly. "Right in the middle of a radio broadcast, sir." "I don't want to listen to your gabbling when I'm hunting!" Extrone tore off the head-set and handed it to the bearer. "If they call back, find out what they want, first. I don't want to be bothered unless it's important." "Yes, sir." Extrone squinted up at the sun; his eyes crinkled under the glare, and perspiration stood in little droplets on the back of his hands. Lin, returning to the column, threaded his way among reclining bearers. He stopped before Extrone and tossed his hair out of his eyes. "I located a spoor," he said, suppressed eagerness in his voice. "About a quarter ahead. It looks fresh." Extrone's eyes lit with passion. Lin's face was red with heat and grimy with sweat. "There were two, I think." "Two?" Extrone grinned, petting the rifle. "You and I better go forward and look at the spoor." Lin said, "We ought to take protection, if you're going, too." Extrone laughed. "This is enough." He gestured with the rifle and stood up. "I wish you had let me bring a gun along, sir," Lin said. "One is enough in my camp." The two of them went forward, alone, into the forest. Extrone moved agilely through the tangle, following Lin closely. When they came to the tracks, heavily pressed into drying mud around a small watering hole, Extrone nodded his head in satisfaction. "This way," Lin said, pointing, and once more the two of them started off. They went a good distance through the forest, Extrone becoming more alert with each additional foot. Finally, Lin stopped him with a restraining hand. "They may be quite a way ahead. Hadn't we ought to bring up the column?" The farn beast, somewhere beyond a ragged clump of bushes, coughed. Extrone clenched the blast rifle convulsively. The farn beast coughed again, more distant this time. "They're moving away," Lin said. "Damn!" Extrone said. "It's a good thing the wind's right, or they'd be coming back, and fast, too." "Eh?" Extrone said. "They charge on scent, sight, or sound. I understand they will track down a man for as long as a day." "Wait," Extrone said, combing his beard. "Wait a minute." "Yes?" "Look," Extrone said. "If that's the case, why do we bother tracking them? Why not make them come to us?" "They're too unpredictable. It wouldn't be safe. I'd rather have surprise on our side." "You don't seem to see what I mean," Extrone said. " We won't be the—ah—the bait." "Oh?" "Let's get back to the column." "Extrone wants to see you," Lin said. Ri twisted at the grass shoot, broke it off, worried and unhappy. "What's he want to see me for?" "I don't know," Lin said curtly. Ri got to his feet. One of his hands reached out, plucked nervously at Lin's bare forearm. "Look," he whispered. "You know him. I have—a little money. If you were able to ... if he wants," Ri gulped, "to do anything to me—I'd pay you, if you could...." "You better come along," Lin said, turning. Ri rubbed his hands along his thighs; he sighed, a tiny sound, ineffectual. He followed Lin beyond an outcropping of shale to where Extrone was seated, petting his rifle. Extrone nodded genially. "The farn beast hunter, eh?" "Yes, sir." Extrone drummed his fingers on the stock of the blast rifle. "Tell me what they look like," he said suddenly. "Well, sir, they're ... uh...." "Pretty frightening?" "No, sir.... Well, in a way, sir." "But you weren't afraid of them, were you?" "No, sir. No, because...." Extrone was smiling innocently. "Good. I want you to do something for me." "I ... I...." Ri glanced nervously at Lin out of the tail of his eye. Lin's face was impassive. "Of course you will," Extrone said genially. "Get me a rope, Lin. A good, long, strong rope." "What are you going to do?" Ri asked, terrified. "Why, I'm going to tie the rope around your waist and stake you out as bait." "No!" "Oh, come now. When the farn beast hears you scream—you can scream, by the way?" Ri swallowed. "We could find a way to make you." There was perspiration trickling down Ri's forehead, a single drop, creeping toward his nose. "You'll be safe," Extrone said, studying his face with amusement. "I'll shoot the animal before it reaches you." Ri gulped for air. "But ... if there should be more than one?" Extrone shrugged. "I—Look, sir. Listen to me." Ri's lips were bloodless and his hands were trembling. "It's not me you want to do this to. It's Mia, sir. He killed a farn beast before I did, sir. And last night—last night, he—" "He what?" Extrone demanded, leaning forward intently. Ri breathed with a gurgling sound. "He said he ought to kill you, sir. That's what he said. I heard him, sir. He said he ought to kill you. He's the one you ought to use for bait. Then if there was an accident, sir, it wouldn't matter, because he said he ought to kill you. I wouldn't...." Extrone said, "Which one is he?" "That one. Right over there." "The one with his back to me?" "Yes, sir. That's him. That's him, sir." Extrone aimed carefully and fired, full charge, then lowered the rifle and said, "Here comes Lin with the rope, I see." Ri was greenish. "You ... you...." Extrone turned to Lin. "Tie one end around his waist." "Wait," Ri begged, fighting off the rope with his hands. "You don't want to use me, sir. Not after I told you.... Please, sir. If anything should happen to me.... Please, sir. Don't do it." "Tie it," Extrone ordered. "No, sir. Please. Oh, please don't, sir." "Tie it," Extrone said inexorably. Lin bent with the rope; his face was colorless. They were at the watering hole—Extrone, Lin, two bearers, and Ri. Since the hole was drying, the left, partially exposed bank was steep toward the muddy water. Upon it was green, new grass, tender-tuffed, half mashed in places by heavy animal treads. It was there that they staked him out, tying the free end of the rope tightly around the base of a scaling tree. "You will scream," Extrone instructed. With his rifle, he pointed across the water hole. "The farn beast will come from this direction, I imagine." Ri was almost slobbering in fear. "Let me hear you scream," Extrone said. Ri moaned weakly. "You'll have to do better than that." Extrone inclined his head toward a bearer, who used something Ri couldn't see. Ri screamed. "See that you keep it up that way," Extrone said. "That's the way I want you to sound." He turned toward Lin. "We can climb this tree, I think." Slowly, aided by the bearers, the two men climbed the tree, bark peeling away from under their rough boots. Ri watched them hopelessly. Once at the crotch, Extrone settled down, holding the rifle at alert. Lin moved to the left, out on the main branch, rested in a smaller crotch. Looking down, Extrone said, "Scream!" Then, to Lin, "You feel the excitement? It's always in the air like this at a hunt." "I feel it," Lin said. Extrone chuckled. "You were with me on Meizque?" "Yes." "That was something, that time." He ran his hand along the stock of the weapon. The sun headed west, veiling itself with trees; a large insect circled Extrone's head. He slapped at it, angry. The forest was quiet, underlined by an occasional piping call, something like a whistle. Ri's screams were shrill, echoing away, shiveringly. Lin sat quiet, hunched. Extrone's eyes narrowed, and he began to pet the gun stock with quick, jerky movements. Lin licked his lips, keeping his eyes on Extrone's face. The sun seemed stuck in the sky, and the heat squeezed against them, sucking at their breath like a vacuum. The insect went away. Still, endless, hopeless, monotonous, Ri screamed. A farn beast coughed, far in the matted forest. Extrone laughed nervously. "He must have heard." "We're lucky to rouse one so fast," Lin said. Extrone dug his boot cleats into the tree, braced himself. "I like this. There's more excitement in waiting like this than in anything I know." Lin nodded. "The waiting, itself, is a lot. The suspense. It's not only the killing that matters." "It's not only the killing," Lin echoed. "You understand?" Extrone said. "How it is to wait, knowing in just a minute something is going to come out of the forest, and you're going to kill it?" "I know," Lin said. "But it's not only the killing. It's the waiting, too." The farn beast coughed again; nearer. "It's a different one," Lin said. "How do you know?" "Hear the lower pitch, the more of a roar?" "Hey!" Extrone shouted. "You, down there. There are two coming. Now let's hear you really scream!" Ri, below, whimpered childishly and began to retreat toward the tether tree, his eyes wide. "There's a lot of satisfaction in fooling them, too," Extrone said. "Making them come to your bait, where you can get at them." He opened his right hand. "Choose your ground, set your trap. Bait it." He snapped his hand into a fist, held the fist up before his eyes, imprisoning the idea. "Spring the trap when the quarry is inside. Clever. That makes the waiting more interesting. Waiting to see if they really will come to your bait." Lin shifted, staring toward the forest. "I've always liked to hunt," Extrone said. "More than anything else, I think." Lin spat toward the ground. "People should hunt because they have to. For food. For safety." "No," Extrone argued. "People should hunt for the love of hunting." "Killing?" "Hunting," Extrone repeated harshly. The farn beast coughed. Another answered. They were very near, and there was a noise of crackling underbrush. "He's good bait," Extrone said. "He's fat enough and he knows how to scream good." Ri had stopped screaming; he was huddled against the tree, fearfully eying the forest across from the watering hole. Extrone began to tremble with excitement. "Here they come!" The forest sprang apart. Extrone bent forward, the gun still across his lap. The farn beast, its tiny eyes red with hate, stepped out on the bank, swinging its head wildly, its nostrils flaring in anger. It coughed. Its mate appeared beside it. Their tails thrashed against the scrubs behind them, rattling leaves. "Shoot!" Lin hissed. "For God's sake, shoot!" "Wait," Extrone said. "Let's see what they do." He had not moved the rifle. He was tense, bent forward, his eyes slitted, his breath beginning to sound like an asthmatic pump. The lead farn beast sighted Ri. It lowered its head. "Look!" Extrone cried excitedly. "Here it comes!" Ri began to scream again. Still Extrone did not lift his blast rifle. He was laughing. Lin waited, frozen, his eyes staring at the farn beast in fascination. The farn beast plunged into the water, which was shallow, and, throwing a sheet of it to either side, headed across toward Ri. "Watch! Watch!" Extrone cried gleefully. And then the aliens sprang their trap. | A. they both think Extrone is going to kill them |
What is the biggest effect when criminals are noted as having similar facial features to other wrongdoers?
A. These sets of criminals are often shown to have similar socioeconomic backgrounds
B. This occurs when people are making judgements but not computers
C. This perpetuates the belief in the area of study that should not be held up
D. These coincidences are held under scrutiny and often disproved
| Face value When the BBC broadcast the recent documentary by Louis Theroux that looked back at the time he spent in the company of Jimmy Savile, there was disbelief across social media that no one had stepped in to stop Savile from committing his crimes. Some blamed the BBC, some blamed those in Savile's immediate circle, but others blamed a simple error of human judgment. "He literally couldn't look more like a paedophile," read one post – one of many to state a supposedly incontrovertible truth: that Savile's criminal tendencies could have been detected from the shape of his features, his eyes, his hair. Moreover, this has nothing to do with the benefit of hindsight and should have been picked up at the time. His looks, they suggested, were a moral indicator, with a wealth of compelling visual evidence to support the claim. We know that paedophiles, murderers and other violent criminals come in many shapes and sizes. If we knew nothing about their criminal history, some of their photos might even appear attractive. But the idea that someone's features betray their character is something rooted deep within us; it's the reason why certain photos perform well on dating apps, or why trustworthy-looking politicians might rack up votes. But how wrong are our hunches of perceived criminality? A recent paper, published by Xiaolin Wu and Xi Zhang of Shanghai's Jiao Tong University, claims to be the first to use machine learning and neural networks to attempt a fully automated inference of criminality from facial images, removing prejudice from the equation and testing the validity of our gut feelings. "What facial features influence the average Joe's impulsive and yet consensual judgments on social attributes?" they ask. Through a study of 1,856 images ("controlled for race, gender, age and facial expression") they claim to have established the validity of "automated, face-induced inference on criminality, despite the historical controversy surrounding this line of enquiry." In other words, they believe that they've found a relationship between looking like a criminal and actually being one. It's a claim that's been made many times over the years. Physiognomy, the 'science' of judging people by their appearance, was first theorised by the ancient Greeks in around the 5th century BC. Aristotle's pronouncement that "it is possible to infer character from features" led to a number of works relating to 'Physiognomica', a word derived from physis (nature), nomos (law) and (or) gnomon (judge or interpreter). All of Greek society, it was claimed, could benefit from this skill: it could assist with choosing an employee, a slave or a spouse, while its inherent vagueness made it intriguing to philosophers and useful for scientists who bent the theories to support their own beliefs. It became a recognised science in the Islamic world, and was used and taught in Europe throughout late antiquity and the early Middle Ages, despite nagging doubts among thinkers and physicians of the day. In the early 16th century, Leonardo da Vinci claimed not to "concern myself with false physiognomy, because these chimeras have no scientific foundation." Theories of physiognomy, however, would persist beyond the Renaissance. In 1586, Italian scholar Giambattista della Porta published a book, De humana physiognomonia libri IIII, which established him as the 'father of Physiognomy'. Della Porta's thinking was based on the 'doctrine of signatures'; the idea that the appearance of plants and animals offers clues to their nature. For example, as one writer of the time suggested, walnuts are good for curing headaches because they're shaped a bit like a human head. The theories in della Porta's book were supported by dozens of detailed illustrations which, by comparing human faces to those of animals, suggested that they must surely share similar character traits. In the 17th century, Swiss poet Johann Caspar Lavater took della Porta's methodology and ran with it, commissioning artists to illustrate his popular Essays On Physiognomy – which, to the chagrin of his contemporary, the writer Hannah More, sold for "fifteen guineas a set… while in vain we boast that philosophy [has] broken down all the strongholds of prejudice, ignorance, and superstition." Lavater's work was criticised for being ridden with bias (black faces rarely emerged well from his analyses) but he was right in one respect: "Whether they are or are not sensible of it," he wrote, "all men are daily influenced by physiognomy." Many studies have been done into our psychological response to faces, and it's clear that a so-called halo effect will inevitably work its magic. "Attractive people are regarded as better at everything," says Professor Peter Hancock, lecturer in Psychology at Stirling University. "And we can't shake that off because there's some truth to it. Good genes produce intelligent people, attractive faces, fit bodies, and we imagine that they're going to be good at everything else, too. We don't have good insight into our own behaviour. We tend to think we understand what we're doing, but we don't." Hancock describes attending a conference where one speaker showed a series of black faces and white faces to students (who were mostly white) and asked them what they thought the experiment was about. "They knew that he was trying to assess whether they would rate the black ones as more criminal," says Hancock. "But then they did!" We attribute social characteristics based on opinions we already hold about certain kinds of faces: whether they look unusual in some way, whether they resemble a partner, a family member or even ourselves, or perhaps have some other cultural association. Physiognomy ultimately stems from what Alexander Todorov, professor of psychology at Princeton University, calls an 'overgeneralisation hypothesis'. "People," he wrote, "use easily accessible facial information (eg an expression such as a smile, cues to gender and ethnic group) to make social attributions congruent with this information (eg a nice person)." In a social media age, the pictures we choose to represent ourselves online are a form of self-presentation driven by those social attributions and the knowledge that our pictures are being judged. Experiments at Princeton found that we take less than one tenth of a second to form an opinion of strangers from their pictures, and those opinions tend to stand firm even if we're exposed to those pictures for a longer period of time. That tendency to judge instantly gives rise to a number of selfie tropes that are deemed to elicit positive responses, particularly when it comes to photos on dating profiles: certain angles, particular expressions, minute adjustments of eyebrows and lips that might appear to be about narcissism and vanity, but are more about a fear of being incorrectly assessed. After all, false suppositions based on people's faces are hugely influential within society, and in extreme cases they can have a huge impact on people's lives. When retired teacher Christopher Jefferies was held by police in connection with the murder of Joanna Yeates in Bristol back in 2010, more than half a dozen newspapers gave his unusual appearance particular scrutiny and made assumptions accordingly, which in turn influenced public opinion. This culminated in substantial damages for defamation, two convictions for contempt of court and a painful ordeal for Jefferies, who was entirely innocent. This kind of deep-seated bias looms large throughout physiognomic works of the 19th and 20th centuries, from absurdities such as Vaught's Practical Character Reader of 1902 (handy if you want to find out what a "deceitful chin" looks like) to more inherently troubling volumes such as Cesare Lombroso's Criminal Man. After performing a number of autopsies on criminals, the Italian physician claimed to have discovered a number of common characteristics, and it's worth listing them if only to establish the supposed criminality of pretty much everyone you know: Unusually short or tall height; small head, but large face; fleshy lips, but thin upper lip; protuberances on head and around ear; wrinkles on forehead and face; large sinus cavities or bumpy face; tattoos; receding hairline; large incisors; bushy eyebrows, tending to meet across nose; large eye sockets but deep-set eyes; beaked or flat nose; strong jaw line; small and sloping forehead; small or weak chin; thin neck; sloping shoulders but large chest; large, protruding ears; long arms; high cheek bones; pointy or snubbed fingers or toes. In a woeful misreading of Darwinian theory, Lombroso unwittingly founded the field of anthropological criminology, and more specifically the idea of the born criminal: a hereditary quality that posed a danger to society and must be rooted out. His theories became discredited during the 20th century, but the kind of bias displayed by Lombroso can still be found in legal systems across the world; studies show that people with stereotypically 'untrustworthy' faces tend to receive harsher treatment than those who don't. There's evidently some consensus over people's attitudes toward certain faces, but it doesn't follow that the consensus is correct. The only attributes that we're reasonably good at detecting, according to research done at the University of Michigan in the 1960s and later tested at the University of Stirling in 2007, are extroversion and conscientiousness. For other traits there's insufficient evidence that our hunches are correct, with anomalies explained by our evolved aversion to 'ugliness', established links between broader faces and powerful physiques, or cultural associations with certain demographics which are reinforced with nagging regularity by newspapers, books, television and film. Data-driven studies, based upon huge quantities of facial data, would seem to offer the final word on this. Since 2005, computational models have used various techniques to test for links between social attributes and facial features, resulting in suggestions that our faces can betray, for example, political leanings, sexual orientation and criminality. One BBC Future article from 2015 even describes the 'discipline' of physiognomy as 'gaining credibility'. But Todorov details many problems with these studies, pointing out the challenging nature of doing such experiments with sufficient rigour – not least because different images of the same people can prompt wildly differing results. The aforementioned study at Shanghai's Jiao Tong University, with its enthusiastic, data-driven analyses of such questions as "What features of a human face betray its owner's propensity for crimes?" prompted a wave of press coverage. The vision outlined in these articles is of an unethical dystopia where neural networks can assess our faces and establish a likely score for criminality – but Todorov is scathing about this paper, too. "The main problem is the sampling of the images," he says. "There is not enough information about the [nature of] the images of the people who were convicted. Second, clearly, there are huge differences between the two samples [of convicts and non-convicts] [in terms of] education and socio-economic status." In other words, your appearance is affected by the kind of life you've led, so the classifiers within the computer program are simply distinguishing between different demographics rather than detecting a propensity for criminal behaviour. Todorov is also wary of these classifiers misidentifying more 'innocent' people than identifying actual criminals, and accuracy is a concern shared by Peter Hancock. "Networks don't assess faces in the same way that we do," he says. "One of our systems, which is a deep network, has a recognition engine which generates an ordered list of how similar various faces are. And sometimes you get good matches – but other times you look at them and say, well, it's the wrong race! To humans they look completely different. And that underlines the fact that the networks are working in a different sort of way, and actually you don't really know how they're working. They're the ultimate black box." This isn't to say that the use of big data, and particularly the use of composite imagery (digitally blending together certain types of faces) doesn't give us useful information and fascinating correlations. "You can, for example, take a given face and use computer software to make it look more or less trustworthy," says Hancock. "I remember a colleague playing with this and he made a less trustworthy version of George W Bush – and how shifty did he look! I'm surprised that they're not using these techniques in political advertising, because you couldn't tell that anything had been done [to the picture], but when you look at it you think 'I wouldn't trust him'." The revitalisation of the theory of physiognomy by the Shanghai students is, according to Todorov, deeply problematic on a theoretical level. "Are we back to Lombroso's theory," he asks, "that criminals were anomalous creatures, evolutionary degenerates? How does one become criminal, and what role do various life forces play into this? There are people making claims that you just need to look at the face to predict personality and behaviour, but many of these people have not given much thought to their underlying assumptions." While it's true that we judge books by their covers, covers are more than just faces; we piece together all kinds of cues from people to form our impressions of them. Jimmy Savile's appearance was unusual by any standards, but we absorbed a great deal of information about him over the years that will have influenced our opinions – not least from the original Louis Theroux programme from 2000 that was reexamined in that recent BBC documentary. Savile's vague resemblance to the Child Catcher from the film Chitty Chitty Bang Bang is convenient but ultimately misleading, and the way it reinforces the idea of what a paedophile might 'look like' is unfortunate; not least because it helps to sustain a low-level belief in the 'science' of physiognomy, despite its tendency to crumble under the slightest cross examination. This article was originally published on TheLong+Short. Read the original article. | C. This perpetuates the belief in the area of study that should not be held up |
What are three possible phases for language formation? | ### Introduction
This letter arises from two intriguing questions about human language. The first question is: To what extent language, and also language evolution, can be viewed as a graph-theoretical problem? Language is an amazing example of a system of interrelated units at different organization scales. Several recent works have stressed indeed the fact that human languages can be viewed language as a (complex) network of interacting parts BIBREF0, BIBREF1, BIBREF2, BIBREF3. Within the graph-based approach to human language, one may think word-meaning mappings (that is, vocabularies) as bipartite graphs, formed by two sets: words and meanings BIBREF2. The second question is: What is the nature of the language evolution process that affects the shape of graph-based language representations? To answer this question, we assume that human communication is constrained (at least) by two forces BIBREF2: one that pushes towards communicative success and another one that faces the trade-off between speaker and hearer efforts. The first force involves simpler decentralized models of linguistic interactions within populations of artificial agents, endowed with minimal human cognitive features, negotiating pieces of a common language: the so-called language games BIBREF4, BIBREF5, BIBREF6, BIBREF7. In the simplest language game, the naming game BIBREF8, BIBREF9, at discrete time step a pair of players (typically one speaker and one hearer) interacts towards agreement on word-meaning associations. Next, we also consider the communication cost to establish word-meaning mappings. G. Zipf referred to the lexical trade-off between two competing pressures, ambiguity and memory, as the least effort principle BIBREF10, BIBREF11: speakers prefer to minimize memory costs; whereas, hearers prefer to minimize disambiguation costs. As remarked by several works, an interesting proposal has stated that human-like vocabularies appear as a phase transition at a critical stage for both competing pressures BIBREF12, BIBREF13, BIBREF14, BIBREF15, BIBREF16. The appearance of a drastic stage of competing pressures can be understood moreover as an explanation of the empirical Zipf's law, which establishes a dichotomy between low-memory words (like the word “the") and low-ambiguity words (like the word “cat"). Within a statistical point of view, text corpora evidence strong scaling properties in word-frequencies BIBREF17, BIBREF18, BIBREF19, BIBREF20, BIBREF21, BIBREF22, BIBREF23, BIBREF24. The main aim is to address a decentralized approach (based on a previous proposal of two authors of this letter BIBREF25) to the emergence of Zipfian properties in a human-like language, while players communicate with each other using bipartite word-meaning mappings. To structurally characterize changes in the system, our methodology is mainly based on a phase transition description, arising from both classical statistical mechanics tools and graph-mining techniques. We run numerical simulations over simple population topologies. We apply graph-mining techniques, particularly a clustering notion for bipartite graphs BIBREF26. ### The model ::: Key concepts on (bipartite) graphs
A bipartite graph is a triple $B=(\top ,\bot ,E)$, where $\top $ and $\bot $ are two mutually disjoint set of nodes, and $E \subseteq \top \times \bot $ is the set of edges of the graph. Here, $\top $ represents the set of word nodes, whereas $\bot $ represents the set of meaning nodes. We remark that edges only exist between word nodes and meaning nodes. A classical useful tool in graph theory is the matrix representation of graphs. Here, we only consider the adjacency matrix. Let us denote by $A = (a)_{wm}$ the adjacency matrix for the (bipartite) graph $B$. From the bipartite sets $\top $ and $\bot $, representing respectively word and meaning nodes, we define the rows of $A$ as word nodes, and the columns as meaning nodes, where $(a)_{wm}=1$ if the word $w$ is joined with the meaning $m$, and 0 otherwise. The neighbors of order 1 of $u\in \top $ are the nodes at distance 1: $N(u)=\lbrace v\in \bot : uv \in E\rbrace $ (if $u \in \bot $ the definition is analogous). Let us denote by $N(N(u))$ the set of nodes at distance 2 from $u$. The degree $d(u)$ of the node $u$ is simply defined by $d(u)=|N(u)|$. We denote by $d^{\max }_W = \max _{w \in W} d(w)$ the maximum degree for word nodes ($\top $). Analogously, $d^{\max }_M = \max _{m \in M} d(m)$ the maximum degree for meaning nodes ($\bot $). The notion of clustering coefficient (in classical graphs) captures the fact that when there is an edge between two nodes they probably have common neighbors. More generally, such notion captures correlations between neighborhoods. Based on this point of view, BIBREF26 proposed a clustering coefficient notion for bipartite graphs: where $cc(u,v)$ is a notion of clustering defined for pairs of nodes (in the same set $\top $ or $\bot $): Interestingly, $cc(u,v)$ captures the overlap between the neighborhoods of $u$ and $v$: if $u$ and $v$ do not share neighbors $cc(u,v)=0$; if they have the same neighborhood $cc(u,v)=1$. To give an overall overview of bipartite clustering for the graph $B$, the average bipartite clustering reads ### The model ::: Basic elements of the language game
The language game is played by a finite population of participants $P=${1,...,p}, sharing both a set of words $W=\lbrace 1,...,n\rbrace $ and a set of meanings $M=\lbrace 1,...,m\rbrace $. Each player $k\in P$ is endowed with a graph-based word-meaning mapping $B^k=(\top ^k,\bot ^k,E^k)$. In our case, $B^k$ is a bipartite graph with two disjoint sets: $\top ^k \subseteq W$ (word nodes) and $\bot ^k \subseteq M$ (meaning nodes). Each player $k \in P$ only knows its own graph $B^k$. Two technical terms are introduced. First, we say that a player $k \in P$ knows the word $w \in W$ if $w \in \top ^k$. Clearly, this definition is equivalent to the existence of the edge $wm \in E^k$, for some $m \in \bot ^k$. Second, the ambiguity of the word $w$, denoted $a(w)$, is defined as its node degree $d(w)$. ### The model ::: Language game rules
The dynamics of the language game is based on pairwise speaker-hearer interactions at discrete time steps. At $t \geqslant 0$, a pair of players is selected uniformly at random: one plays the role of speaker $s$ and the other plays the role of hearer $h$, where $s,h \in P$. Each speaker-hearer communicative interaction is defined by two successive steps. The speaker-centered STEP 1 involves the selection of a meaning and a word to transmit them. At STEP 2, the hearer receives the word-meaning association and both speaker and hearer behave according to either repair or alignment strategies. STEP 1. To start the communicative interaction, the speaker $s$ selects the topic of the conversation: one meaning $m^* \in M$. To transmit the meaning $m^*$, the speaker needs to choose some word, denoted $w^*$. There are two possibilities for the selection of $w^*$: if the edge $wm^* \notin E^s$ for any $w \in \top ^s$, the speaker chooses (uniformly at random) the word $w^*$ from the set $W$ and adds the edge $w^*m^*$ to the graph $B^s$; otherwise, if $w^*m^* \in E^s$ for some $w^* \in \top ^s$, the speaker calculates $w^*$ based on its interests, that is, based on its own conflict between ambiguity and memory. To calculate $w^*$ for the second case ($w^*m^* \in E^s$), the speaker behaves according to the ambiguity parameter $\wp \in [0,1]$. Let $random \in [0,1]$ be a random number. Then, two actions are possible: if $random \geqslant \wp $, the speaker calculates $w^*$ as the least ambiguous word otherwise, the speaker calculates $w^*$ as the most ambiguous word The speaker transmits the word $w^*$ to the hearer. STEP 2. The hearer behaves as in the naming game. On the one hand, mutual speaker-hearer agreement (if the hearer knows the word $w^*$) involves alignment strategies BIBREF9. On the other hand, a speaker-hearer disagreement (if the hearer does not know the word $w^*$) involves a repair strategy in order to increase the chance of future agreements (that is, for $t^{\prime }>t$). More precisely, if the hearer knows the word $w^*$, both speaker and hearer remove all edges formed by $wm^*$, where $w$ respectively belongs to $\top ^s \setminus \lbrace w^*\rbrace $ and $\top ^h \setminus \lbrace w^*\rbrace $. otherwise, the hearer adds the edge $w^*m^*$ to its graph $B^h$. ### Methods
The population of agents is located on the vertices of a complete graph of size $|P|=100$, typically called the mean field approximation. For the description of other simple graph topologies, see the caption of Fig. FIGREF15. The population shares both a set of $n=|W|=128$ words and a set of $m=|M|=128$ meanings. Starting from an initial condition in which each player $k \in P$ is associated to a bipartite graph $B^k$ where $B^k_{ij} = 1$ or $B^k_{ij} = 0$ with probability 0.5 (put differently, for each possible edge $ij$, $i \in W$ and $j \in M$, exists with probability 0.5), the dynamics performs a speaker-hearer interaction at each discrete time step $t \geqslant 0$. The bipartite word-meaning mappings $B^s$ and $B^h$ are then reevaluated according to communicative success. All results consider averages over 10 initial conditions and $3\times 10^5$ time steps. We denote by $t_f$ the final time step. The ambiguity parameter $\wp $ is varied from 0 to 1 with an increment of 1%. ### Results ::: Three structural phases in language formation
Two key quantities have been analyzed for different values of $\wp $: the average population clustering $cc$, which captures the average correlation between word neighborhoods; and the (effective) lexicon size at time step $t$, $V(t)$, defined as BIBREF12, BIBREF25 where $V(t)=1$ if $|\top ^k|=n$, while $V(t)=0$ if $|\top ^k|=0$. Three clear domains can be noticed in the behavior of $\langle cc \rangle $ versus $\wp $, at $t_f$, as shown in Fig. FIGREF15 (blue squares). Phase I: $\langle cc \rangle $ increases smoothly for $\wp < 0.4$, indicating that for this domain there is a small correlation between word neighborhoods. Full vocabularies are attained also for $\wp < 0.4$; Phase II: a drastic transition appears at the critical domain $\wp ^* \in (0.4,0.6)$, in which $\langle cc \rangle $ shifts abruptly towards 1. An abrupt change in $V(t_f)$ versus $\wp $ is also found (Fig. FIGREF16) for $\wp ^*$; Phase III: single-word languages dominate for $\wp > 0.6$. The maximum value of $\langle cc \rangle $ indicate that word neighborhoods are completely correlated. ### Results ::: Bipartite graphs to visualize the phase transition
We now shift our focus from graph-based measures towards a holistic level in which we illustrate the described phase transition using bipartite graph representations of language formation. We stress the fact that our framework based on a language game with players endowed with bipartite word-meaning mappings is able to visualize the structural changes of the three phases (I, II and III). Fig. FIGREF18 display, from top to bottom, the bipartite word-meaning mappings for ambiguity parameters $\wp $ in $\lbrace 0.1, 0.52,1\rbrace $. As expected, there are radical structural changes between bipartite graphs associated to such ambiguity parameters. Full vocabularies are attained for $\wp =0.1$ (Phase I), located at the hearer-centered phase. Zipfian vocabularies seem to appear for $\wp =0.52$ (Phase II), where speaker and hearer costs have a similar value. Finally, a single-word vocabulary (that is, one word, several meanings) is exhibited for $\wp =1$ (Phase III). ### Results ::: Critical values of energy
The appearance of the three-phased language behavior described here is closely related to previous results of two authors of this letter BIBREF25. Indeed, in the cited paper the energy-like functional $e_{KL}$ (a kullback-leibler-based measure) is minimized around the parameter $\wp \approx 0.5$. Remarkably, here it is showed numerically that around the critical parameter $\wp \approx 0.52$ a drastic transition for both the effective vocabulary and the bipartite average clustering tends to appear (see Fig. FIGREF18). A first strategy to profound on the problem established between the phase transitions described here and energy-based approaches, is to measure the information-theoretic energy $\Omega _\wp (tf)$ (as defined in BIBREF12) as a function of the parameter $\wp $. $\Omega _\wp (tf)$ is a combination of the respective efforts of speakers and hearers: $\Omega _\wp (tf)=\wp H(R|S)+(1-\wp )H(S)$. Figure FIGREF20 showed that $\Omega _\wp (t_f)$ is minimized around $\wp \approx 0.5$. This suggests a new way to understand language evolution and formation, by reconciling models focused on self-organization and information-theoretic accounts. ### Discussion
In this letter, we have described a decentralized model of the emergence of Zipfian features in a human-like language, where agents play language games communicating with bipartite word-meaning mappings. The model evidences a phase transition that corresponds to the formation of a human-like vocabulary satisfying Zipfian word-meaning properties. Our central graph-mining tool has been a notion of clustering for bipartite graphs. This function allowed us to suggest that the drastic transition is, in some sense, a qualitative transition in word's correlations. To further understand the nature of the described transition, we remark a recent proposal BIBREF28, reinterpreting an old question about language learning with a novel approach: if language learning by a child involves setting many parameters, to what extent all these need to be innate? According to the Principles and Parameters theory BIBREF29, children are biologically endowed with a general “grammar" and then the simple exposition to a particular language (for example, Quechua) fixes its syntax by equalizing parameters. This debate was illuminated by proposing a statistical mechanics approach in which the distribution of grammar weights (where language is modeled by weighted context-free grammars) evidences a drastic transition. Language learning is, for this proposal, a transition from a random model of grammar parameter-weights to the one in which deep structure (that is, syntax) is encountered. Here, the language learning problem is situated in a decentralized process, with agents negotiating a common word-meaning mapping exhibiting Zipfian scaling properties. Interestingly, our approach can shed light on the debate opened by BIBREF28. Indeed, our model questioned, first, the fact that language learning is traditionally viewed as an individual process, without any consideration of population structure (in general, language games question this fact). Secondly, we argue that our view pointed out the minimal necessity of cognitive principles for cultural language formation: the least effort principle. We hypothesize that players only need the most basic cognitive features for language learning (and formation) and the rest is an emergent property from the local speaker-hearer interactions. It is interesting to remark that several works have stressed the fact that language formation can be viewed as a phase transition within an information-theoretic approach BIBREF12, BIBREF13, BIBREF14, BIBREF15, BIBREF16. Future work could explore an intriguing hypothesis: Zipfian properties have strong consequences for syntax and symbolic reference. BIBREF30 has proposed indeed that Zipf's law is a necessary precondition for full syntax, and for going beyond simple word-meaning mappings. They hypothesized moreover that the appearance of syntax have been as abrupt as the transition to Zipf's law. This is a goal for future work: to propose a decentralized model in which agents (constrained by specific cognitive features) develop a Zipfian language that acts as a precondition for the abrupt transition to simple forms of syntax (based, for example, on BIBREF31). Another related research line arises from models assuming the interplay between maximization of the information transfer and minimization of the entropy of signals (see, for example, BIBREF32). As previously remarked, these models evidence a lack of population structure. Current work asks how a community of individuals playing the language game proposed here can minimize the energy functional $\Omega (\lambda )$ of word-meaning mappings. We may hypothesize that reaching global consensus at the critical phase (that, is for $\wp \approx 0.5$) is closely related to the global minima of $\Omega (\lambda )$ (as a first approach to this problem, see Section SECREF19). This idea opens fascinating novel ways to study human language, reconciling models seeing communication as a global minima of information entropic energies and models focused on populations self-organizing themselves towards a shared consensus. F.U. thanks CONICYT Chile for financial support under the Grant 3180227. Figure 1: Average population clustering 〈cc〉 versus ambiguity parameter ℘. It is described the behavior of the average population clustering versus ℘. 〈cc〉 denotes the average over 10 realizations. As shown in the figure, three phases for the evolution of bipartite word-meaning mappings tend to appear: full vocabularies (Phase I), human-like (Phase II) and single-word vocabularies (Phase III). Blue squares indicate the mean-field approximation (denoted MF ). The other two curves indicate a ring topology in which individuals interact at some specific radius r. For r = 2 and r = 50, the three-phased division of human-like language tends to disappear. This fact suggests that our results are strongly affected by topology features. Figure 2: Effective vocabulary V (tf ) versus ambiguity parameter ℘. It is described the average behavior of 〈V 〉 versus ℘. Blue squares indicate the mean-field approximation (MF). Three phases are exhibited for language formation: full vocabularies (Phase I), human-like (Phase II) and single-word vocabularies (Phase III). Figure 3: Visualization of bipartite word-meaning mappings. Orange circles indicate meanings, whereas blue cruxes indicate words. From top to bottom, we choose three bipartite graphs corresponding respectively to ℘ in {0.1, 0.52, 1}. Node positions are based on a Python (https://networkx.github.io/) implementation of the Fruchterman-Reingold algorithm [28]. Figure 4: Ω℘(tf ) versus ambiguity parameter ℘. The figure displays the value of the information-theoretic energy Ω℘(tf ) for one initial condition, after tf = 3× 105 speaker-hearer interactions. ℘ is varied with an increment of 1%. | Phase I: $\langle cc \rangle $ increases smoothly for $\wp < 0.4$, indicating that for this domain there is a small correlation between word neighborhoods. Full vocabularies are attained also for $\wp < 0.4$, Phase II: a drastic transition appears at the critical domain $\wp ^* \in (0.4,0.6)$, in which $\langle cc \rangle $ shifts abruptly towards 1. An abrupt change in $V(t_f)$ versus $\wp $ is also found (Fig. FIGREF16) for $\wp ^*$, Phase III: single-word languages dominate for $\wp > 0.6$. The maximum value of $\langle cc \rangle $ indicate that word neighborhoods are completely correlated |
Among all of the derivative instruments that Verizon used to manage the exposure to fluctuations of foreign currencies exchange rates or interest rates, which one had the highest notional value in FY 2021? | Evidence 0:
Derivative Instruments
We enter into derivative transactions primarily to manage our exposure to fluctuations in foreign currency exchange rates and interest rates.
We employ risk management strategies, which may include the use of a variety of derivatives including interest rate swaps, cross currency
swaps, forward starting interest rate swaps, treasury rate locks, interest rate caps, swaptions and foreign exchange forwards. We do not hold
derivatives for trading purposes.
The following table sets forth the notional amounts of our outstanding derivative instruments:
(dollars in millions)
At December 31,
2021
2020
Interest rate swaps
$
19,779
$
17,768
Cross currency swaps
32,502
26,288
Forward starting interest rate swaps
1,000
2,000
Foreign exchange forwards
932
1,405 | Cross currency swaps. Its notional value was $32,502 million. |
What isn't something mentioned in multiple events?
A. famous people
B. politics
C. technology
D. world events
| Eleven-Twelfths of 1999 In Review When Chatterbox invited readers to nominate events, significant deaths, good and bad movies, etc., for 1999--a year likely to get little attention in the coming weeks, as news organizations choose instead to review the entire century or millennium--the response was overwhelming. Chatterbox had promised to publish his official "1999 In Review" item before Thanksgiving, but some distant memory of a scruple persuaded him to wait till November was over. Nothing ever happens in December. OK, that's not quite true. Hordes of protesters in Seattle are making the World Trade Organization's meeting there a much more exciting TV story than anyone expected it to be. Reader Dan Crist (who finds Chatterbox's habit of referring to himself in the third person "rather annoying and less than professional") points out that Japan bombed Pearl Harbor in Dec. 1941. Also, Chatterbox (moonlighting as "Today's Papers" columnist) observed not quite one year ago that the House of Representatives cast its second presidential-impeachment vote in U.S. history on Dec. 19, 1998. (That same news-filled day, the U.S. ended an air war against Iraq and Bob Livingston said he'd decided not to become House speaker after all.) Two months after the impeachment vote, the Senate failed to convict the president--a highly significant event of 1999 that, for some bizarre reason, slipped Chatterbox's mind until several indignant readers wrote in to remind him of it. By now, it should be clear that Chatterbox isn't much good at year-in-review journalism. Fortunately, Chatterbox's readers are very good at it. He will now turn this survey over to them. ( Disclaimer: Although Chatterbox previously stated that he wouldn't include opinions he disagreed with, that standard proved too confining. Where Chatterbox has solid information or opinions to the contrary, he occasionally interjects below. Obviously stupid or unnecessarily sour reader comments were discarded, but if you don't find your nominee below it doesn't necessarily mean that it was obviously stupid or unnecessarily sour. ) Here are 20 important things that happened in 1999: 1. Most Hated Celebrity--Ever? The New York Times reported on Nov. 10, 1999, that a new record had been set in the latest Times /CBS poll: [Its] highest negative rating ever scored by a person in the news. The honor went to Reform Party candidate Donald Trump, who managed to make an unfavorable impression upon some 70 percent of those polled. The paper noted that this achievement far eclipsed the last comparably negative rating--the 55 percent score attained by Linda Tripp. Presumably this came as no surprise to Mr. Trump, who, upon announcing the formation of a presidential exploratory committee on Oct. 7, 1999, had cited polls with "amazing results"--a remark that was widely misinterpreted at the time. -- Jodie Allen of U.S. News & World Report (and frequent Slate contributor) 2. Most Foolishly Ignored Parts of the World in 1999 The dog that did bark but no one noticed--the political turmoil in the three great South Asian nations of India, Pakistan, and Indonesia, which now are well on the way to passing the three northern Asian nations of China, Japan, and Russia in population (Indonesia is fourth, Pakistan just passed Japan to seventh, India will soon pass China to first). But Americans are still fixated on northern Asia--Clinton says he must deal with China, because "you can't ignore a billion people with nuclear weapons," but his own policy toward India shows that you sure can! --Jim Chapin 3. Worst/Best Films of 1999 Here's my nominee for worst movie of the year (complete category should be: "Worst Movie of the Year That Assumedly Adult Male Reviewers Slathered Over"): There's Something About Mary --a pathetically sophomoric, penis-obsessed mess that wouldn't even appeal to Larry Flynt! -- Felicia, Menlo Park, Cal. Chatterbox replies: You've got the wrong year. That was 1998 . [Chatterbox didn't have the heart to add that he thought There's Something About Mary was pretty funny, especially the joke about "the franks or the beans."] Felicia replies: Oops ... well then, the best of '99 was The Red Violin --lyrical, magical, musical, wonderful! [Chatterbox hasn't seen it.] 4. Most Shameless (and Unsuccessful) Attempt To Have It Both Ways in 1999 : Sen. Arlen Specter, citing Scottish law, finds Clinton "not proven" on the impeachment charges. --Andrew Solovay 5. Rest in Peace in 1999: Stanley Kubrick (multiple sources) John Kennedy Jr. (multiple sources) Susan Strasberg (anonymous tipster; Strasberg played Anne Frank in the original production of the Broadway adaptation, which some people think wasn't Jewish enough) Mel Torme (Steve Reiness) Mrs. Whozit [ Chatterbox interjects : her name was Anne Sheafe Miller], the first person ever to be saved by penicillin (Blair Bolles) 6. 1999: The Road Not Taken What an extraordinary year! A right-wing conspiracy topples the president, and the governor of Texas reveals himself in a series of debates to be a natural leader with an innate gift for connecting with his audience, a sure sign of his electoral success next year. A new Thomas Harris book brilliantly takes us deeper into the mind of a serial killer; a new Star Wars movie redefines the very nature of entertainment; a new Stanley Kubrick film changes the whole national dialogue about sex and marriage; a new TV series from the creator of SportsNight --oh, I can't even bring myself to bash that piece of do-gooder twaddle. If only McDonald's had come out with three more boldly adult-flavored hamburgers, it would have been a perfect year for dud megaevents--all leading up of course to Y2K, the limpest milestone in human history. --Mike Gebert 7 . Children Behaving Badly in 1999 Don't forget Woodstock 1999 --the concert of "peace and love" that ended in a literal blaze of glory when in an hours-long tribute to the original Woodstock, the mob started ripping down vendor booths and anything else that would burn and piling it onto the bonfires scattered about the scene. [ Chatterbox interjects: Didn't people get assaulted and raped, too?] I'm getting all sentimental just thinking about it. You also left out all the shooting rampages . Several were done in the name of God or love supposedly. They were all committed by "quiet, shy" people who "mostly kept to" themselves. I've started to hang around only loud, obnoxious people. --Susan Hoechstetter 8. A Lunatic Rhapsody for the New York Yankees The Yankees can actually be referred to as the glue that held the century together. Of course, as the 1999 World Series champions, they are a significant "story of the year." However, this one singular achievement must be considered in a broader context. 1999 represented the team's 25th championship of the century. This beats, by one, the most championships any one team won during the century. The Montreal Canadiens have won 23 Stanley Cups. However, the Yankees, an American team, playing in the "City of the Century" (so called by me to reflect the amazing growth and transformation of one city during this period), who play the "National Pastime," are truly an amazing story. The team's first championship occurred in 1921; therefore, they have won 25 of the last 78 years, nearly one in three. This level of sustained excellence is not matched in sports or in any other aspect of society. The 1999 win is possibly the most unique. With free-agency, expansion, and three levels of playoffs, it is much harder to win today than in past years. In fact, by winning three of the last four championships, they are the first team to accomplish this feat during the eras of free-agency and of divisional play. The Sultan of Swat, the Iron Man, the Yankee Clipper, the Mick, and Yogi--these strong, masculine names are synonymous with the team, the sport, and American history. They went hand in hand with two world wars, Superman, and America's superpower status. The 1999 squad does not feature "a name." This team, with its myriad of human-interest stories, its international roster, and no star, is representative of '90s man, male sensitivity, Pax American interests, and the new political paradigm. --Jim Landau from North Potomac, Md. (formerly of the Bronx) 9. A Big Shot Calls for Decriminalizing Drug Use in 1999 New Mexico Gov. Gary Johnson came out for ending drug prohibition. Though this by itself has no immediate effect, it makes it respectable, for the first time, for political leaders to discuss the subject, and thereby brings closer the day when the vast majority of crimes will no longer be committed, when billions of dollars will be freed to help the inner city instead of to ruin black people's lives, and when we will stop, as in Samuel Butler's Erewhon , imprisoning people for the crime of being sick. --Henry Cohen Chatterbox interjects: Didn't Baltimore Mayor Kurt Schmoke do the same thing 11 years ago? 10. Don't Worry in 1999 The Dalai Lama proclaimed that most important thing in the world is to be happy. --Margaret Taylor 11. The Athletic Bra Seen 'Round the World in 1999 Public interest and media attention to the women's World Cup in soccer. --Tom Horton 12. Another Overlooked Foreign-Policy Event in 1999 Presidential primary elections for the first time ever in Mexico. --Tom Horton 13. Policing the World Is Shown To Work in 1999 I nominate as the most under-reported story of the year (and the last few years) the continuing alarmist predictions by foreign-policy and military experts about peacekeeping efforts, which are then proved wrong and immediately forgotten. This year, the obvious one is Kosovo, but the year is also ending with East Timor, where the Aussies and their allies successfully stopped the slaughter with no casualties. These followed Haiti, Bosnia, and Rwanda as places where the West delayed sending in troops because of alarmist predictions. --Jerry Skurnik 14. Barbara Walters Did This One on Her Year-End Special, But It's Still Good Don't forget, Susan Lucci finally won an Emmy . --anonymous tipster 15. Annals of Justice in 1999 Matthew Shepard: the despicable defense . -- anonymous tipster 16. Get Me a New Century, Quick A sitting president was accused of rape. --Ananda Gupta Chatterbox interjects: Yes, but the evidence was shaky--something the Wall Street Journal 's editorial page, which broke the story, was not very forthcoming about. As Jack Shafer wrote in this column, Ronald Reagan, after he left office, was also accused of having once committed rape. The evidence there was shaky, too. 17. The Most Important Thing of All That Happened in 1999 In 1999, more than half of U.S. homes had a PC, for the first time (i.e., home-PC penetration passed 50 percent). Of course, most of these PCs crashed all the time, but it's still a significant development. By the way, Internet hookups in homes are still well below 50 percent. --Walt Mossberg, "Personal Technology" columnist for the Wall Street Journal (and occasional rock-music historian for this column) 18. All Dolled Up and Nowhere To Go in 1999 General Pinochet --Jodie Maurer 19. Senate Endorses Nuclear Proliferation in 1999 The Senate rejected the Comprehensive Test Ban Treaty , thereby decapitating nuclear-arms control and sending Iraq, Iran, and North Korea the message that the United States won't raise a big stink if they try to join India and Pakistan. The president woke up to this possibility at about the moment it was realized, and started lobbying for passage of the treaty a day after it became too late. --Josh Pollack 20. Unremarked Natural Disaster in 1999 The Indian Supercyclone is the biggest, this century at least. --Samir Raiyani Photographs of: Donald Trump by Peter Morgan/Reuters; Natalie Portman by Keith Hamshere/Lucasfilm Ltd./Reuters; New York Yankees players by Gary Hershorn/Reuters; KLA member by Hazir Reka/Reuters. | C. technology |
Are Best Buy's gross margins historically consistent (not fluctuating more than roughly 2% each year)? If gross margins are not a relevant metric for a company like this, then please state that and explain why. | Evidence 0:
Consolidated Statements of Earnings
$ and shares in millions, except per share amounts
Fiscal Years Ended
January 28, 2023
January 29, 2022
January 30, 2021
Revenue
$
46,298
$
51,761
$
47,262
Cost of sales
36,386
40,121
36,689
Gross profit
9,912
11,640
10,573
Selling, general and administrative expenses
7,970
8,635
7,928
Restructuring charges
147
(34)
254
Operating income
1,795
3,039
2,391
Other income (expense):
Investment income and other
28
10
38
Interest expense
(35)
(25)
(52)
Earnings before income tax expense and equity in income of affiliates
1,788
3,024
2,377
Income tax expense
370
574
579
Equity in income of affiliates
1
4
-
Net earnings
$
1,419
$
2,454
$
1,798 | Yes, the margins have been consistent, there has been a minor decline of 1.1% in gross margins between FY2022 and FY2023. |
On clinical neurological assessment, Mrs. Sample demonstrated:
Choose the correct answer from the following options:
A. Intense and persistent spasms in both legs
B. Mild paraparesis in the right leg
C. Severe neural abnormalities
D. Mild paraparesis in the left leg
E. Complete paralysis of the legs
| ### Patient Report 0
**Dear colleague, **
We wish to provide an update regarding Mrs. Anna Sample, born on
01.01.1970. She was admitted to our clinic from 01/01/2017 to
01/02/2017.
**Diagnosis:** Diffuse large B-cell lymphoma of germinal center type; ID
01/2017
- Ann-Arbor: Stage IV
- R-IPI: 2 (LDH, stage)
- CNS-IPI: 2
- Histology: Aggressive B-NHL (DLBCL, NOS); no evidence of t(14;18)
translocation. Ki-67 at 40%. Positive reaction to MUM1, numerous
CD68-positive macrophages. Negative reaction to ALK1 and TdT.
- cMRI: Chronic inflammatory lesions suggestive of Multiple Sclerosis (MS)
- CSF: no evidence of malignancy
- Bone marrow aspiration: no infiltration from the pre-existing
lymphoma.
**Current treatment: **
Initiated R-Pola-CHP regimen q21
- Polatuzumab vedotin: 1.8mg/kg on Day 1.
- Rituximab: 375mg/m² on Day 0.
- Cyclophosphamide: 750mg/m² on Day 1.
- Doxorubicin: 50mg/m on Day 1.
- Prednisone: 100mg orally from Day 1-5.
**Previous therapy and course**
From 12/01/2016: Discomfort in the dorsal calf and thoracic spine,
weakness in the arms with limited ability to lift and grasp, occasional
dizziness.
12/19/2016 cMRI: chronic inflammatory marks indicative of MS.
12/20/2016 MRI: thoracic/lumbar spinal cord: Indication of a
metastatic mass starting from the left pedicle T1 with a significant
extraosseous tumor element and full spinal narrowing at the level of
T10-L1 with pressure on the myelon and growth into the neuroforamen
T11/T12 on the right and T12/L1 left. More lesions suggestive for
metastasis are L2, L3, and L4, once more with extraosseous tumor element
and invasion of the left pedicle.
12/21/2016 Fixed dorsal support T8-9 to L3-4. Decompression via
laminectomy T10 and partial laminectomy lumbar vertebra 3.
12/24/2016 CT chest/abdomen/pelvis: Magnified left axillary lymph node.
In the ventral left upper lobe, indication of a round, loose, cloudy
deposit, i.e., of inflammatory origin, follow-up in 5-7 weeks.
Nodule-like deposit in the upper inner quadrant of the right breast,
senological examination suggested.
**Pathology**: Aggressive B-NHL (DLBCL, NOS); no evidence of t(14;18)
translocation. Ki-67 staining was at 40%. Positive reaction to MUM1.
Numerous CD68-positive macrophages. No reaction to ALK1 and TdT.
**Other diagnoses**
- Primary progressive type of multiple sclerosis (ID 03/02)
- Mood disorder.
- 2-vessel CHD
**Medical History**
Mrs. Sample was transferred inpatient from DC for the initiation of
chemotherapy (R-Pola-CHP) for her DLBCL. In the context of her
pre-existing ALS, she presented on 12/19/2016 with acute pains and
restricted mobility in her upper limbs. After her admission to HK
Flowermoon, an MRI was performed which revealed a thoracic neoplastic
growth especially at the level of T10-L1, but also affecting lumbar
vertebra 3, L4 and L6. Surgical intervention on 12/21/2016 at DC
resulted in symptom relief. Presently, her complaints are restricted to
post-operative spine discomfort, shoulder hypoesthesia, and intermittent
hand numbness. She reported a weight loss of 5 kg during her
hospitalization. She denied having respiratory symptoms, infections,
systemic symptoms, or gastrointestinal complaints. Mrs. Sample currently
has a urinary catheter in place.
**Physical examination on admission**
General: The patient has a satisfactory nutritional status, normal
weight, and is dependent on a walker. Her functional status is evaluated
as ECOG 2. Cardiovascular: Regular heart rhythm at a normal rate. Heart
sounds are clear with no detected murmurs. Respiratory: Normal alveolar
breath sounds. No wheezing, stridor, or other abnormal sounds.
Abdominal: The abdomen is soft, non-tender, and non-distended with
normal bowel sounds in all quadrants. There is no palpable enlargement
of the liver or spleen, and the kidneys are not palpable.
Musculoskeletal: Tenderness noted in the cervical and thoracic spine
area, but no other remarkable findings. This is consistent with her
post-operative status. Lymphatic: No enlargement detected in the
temporal, occipital, cubital, or popliteal lymph nodes. Oral: The oral
mucosa is moist and well-perfused. The oropharynx is unremarkable, and
the tongue appears normal. Peripheral Vascular: Pulses in the hands are
strong and regular. No edema observed. Neurological: Cranial nerves are
intact. There is numbness in both hands and mild hypoesthesia in the
shoulders. Motor strength is 3/5 in the right arm, attributed to her
known ALS diagnosis. No other motor or sensory deficits noted.
Occasional bladder incontinence and intermittent gastrointestinal
disturbances are reported.
**Medications on admission**
Acetylsalicylic acid (Aspirin®) 100 mg: Take 1 tablet in the morning.
Atorvastatin (Lipitor®) 40 mg: Take 1 tablet in the evening. Fingolimod
(Gilenya®) 0.5 mg: Take 1 capsule in the evening. Sertraline (Zoloft®)
50 mg: Take 2 tablets in the morning. Hydromorphone (Dilaudid® or
Exalgo® for extended-release) 2 mg: Take 1 capsule in the morning and 1
in the evening. Lorazepam (Ativan®) 1 mg: 1 tablet as needed.
**Radiology/Nuclear Medicine**
**MR Head 3D unenhanced + contrast from 12/19/2016 10:30 AM**
**Technique:** Sequences obtained include 3D FLAIR, 3D DIR, 3D T2, SWI,
DTI/DWI, plain MPRAGE, and post-contrast MPRAGE. All images are of good
quality. Imaging area: Brain.
There are 20 FLAIR hyperintense lesions in the brain parenchyma,
specifically located periventricularly and in the cortical/juxtacortical
regions (right and left frontal, left temporal, and right and left
insular). No contrast-enhancing lesions are identified. There are also
subcortical/nonspecific lesions present, with some lesions appearing
confluent. The spinal cord is visualized up to the C4 level. No spinal
lesions are noted.
[Incidental findings:]{.underline}
- Brain volume assessment: no indication of reduced brain volume.
- CSF space: age-appropriate usual wide, moderate, and symmetric CSF
spacing with no signs of CSF flow abnormalities.
- Cortical-Subcortical Differentiation: Clear cortical-subcortical
distinction.
- RML-characteristic alterations: none detected.
- Eye socket: appears normal.
- Nasal cavities: Symmetric mucosal thickening with a focus on the
right ethmoidal sinus.
- Pituitary and peri-auricular region: no abnormalities.
- Subcutaneous lesion measuring 14.4 x 21.3 mm, right parietal likely
representing an inflamed cyst or abscess, differential includes soft
tissue growth.
[Evaluation]{.underline}
Dissemination: MRI standards for spatial distribution are satisfied. MRI
criteria for temporal distribution are unfulfilled. Comprehensive
neurological review: The findings are consistent with a chronic
inflammatory CNS disease in the sense of Multiple Sclerosis.
**MR Spine plain + post-contrast from 12/20/2016 10:00 AM**
**Technique:** GE 3T MRI Scanner
MRI was conducted under anesthesia due to claustrophobia.
**Sequences**: Holospinal T2 Dixon sagittal, T1 pre-contrast, T1 fs
post-contrast. The spine is visualized from the craniocervical junction
to S2.
**Thoracic spine: **
On T2-STIR and T2, there is a hyperintense signal of vertebral bodies T5
and T6 with inconsistent delineation of the vertebral endplates,
indicative of age-related changes. There is a reduction in the height of
the disc spaces T4/5 and T5/6 with subligamentous disc protrusion
leading to a spongy appearance of the spinal cord at this location.
Myelon atrophy is noted at T5/6, along with a T2 bright lesion
suggestive of MS at the level of T3 and also T4/5. Spine: A large
intraspinal mass extends from T10-L1, causing an anterior spongy
appearance of the spinal cord and resulting in complete spinal canal
stenosis at this level. On fat-only imaging, there is almost total
replacement of the marrow space of vertebral body T11 with external
tumor extension and infiltration into the lateral structures (more on
the left than the right) and neural foramen T11 on both sides. There is
mild disc herniation at T8/9 with slight sponginess of the spinal cord.
MS-characteristic spinal cord lesions are noted at segments T5 and T8/9.
**Lumbar spine: **
T2-DIXON shows bright signal intensity of the anterior part of lumbar
vertebra 1, a patchy appearance of lumbar vertebra 2, and lumbar
vertebra 4. Almost the entire marrow space is replaced in the fat-only
imaging. There is an external tumor mass posterior to lumbar vertebra 4
without significant spinal canal stenosis, which involves the left
lateral structure and a pronounced appearance of the cauda equina at
lumbar vertebra 1. A call to communicate the results was made at 11:15
a.m. to the on-duty orthopedic surgeon and to colleagues in neurology.
Evaluation Evidence of a metastatic lesion originating from the left
pedicle of T10 with a significant extramedullary tumor mass and full
spinal canal narrowing at the level of T10 with compression of the
spinal cord and extension into the neural foramen T11-T12 on the right,
and T12-L1 on the left. Additional sites suggestive of metastasis
include L2, L3, and L4, again with extramedullary tumor components and
invasion of the left lateral structure. Contrast enhancement of the
distal cord is noted. There are MS-characteristic spinal cord lesions at
the levels of T3, T4-5, T5, and T8-9. The conus medullaris is not
visualized due to spinal cord displacement.
**CT Thoracic Spine from 01/03/2017**
[Clinical Findings]{.underline}
Lateral and medial alignment is stable. No sign of vertebral column
damage. Multiple segment degenerative alterations in the spine. No
indications of mineralization in the recognized space at the level of
T10/L2. Invasion of T10 and L4 with composite osteolytic-osteoblastic
defects starting from the left pedicle into the vertebral column. More
cortical inconsistency with enhanced sclerosis at the endplate of L2.
Review with prior MRI indicative of a different composite defect. Defect
pit at the endplate of lumbar vertebra 2.
Minor pericardial effusion with nearby superior ventilation. Intubation
tube placed. Mild cardiomediastinum. Splenomegaly. Standard display of
the tissue organs of the mid-abdomen, as naturally observed. Normal
spleen. Thin adrenals. Tightly raised kidney bowls and leading ureter
from both aspects, e.g., upon entry during the exhalation period after
gadolinium inclusion in the earlier MRI. No bowel obstruction.
Intestinal stasis. No sign of abnormally magnified lymphatic vessels.
Remaining pin holes in the femoral head on both sides.
[Evaluation]{.underline}
Composite osteolytic-osteoblastic defects starting from the left pedicle
in T10 and lumbar vertebra 4, and at the endplate of lumbar vertebra 2.
**CT Thoracic Spine from 01/04/2017 **
Intraoperative CT imaging for enhanced guidance.
Two intraoperative CT scans were undertaken in total.
On the concluding CT scan, recently implanted non-radiopaque pedicle
screws T8-9 to L2-L3 at tumor band T10. Regular screw placement. No
evident sign of material breakage.
Apart from this, no notable alteration in findings from CT of
01/03/2017.
Evaluation Intraoperative CT imaging for better guidance. Recently
inserted pedicle screws T8/T9 and L2/L3 in tumor indication T10,
ultimate standard screw positioning done transpedicular.
**CT Chest/Abdomen/Pelvis + Contrast from 01/09/2017**
Results: After uneventful intravenous administration of Omnipaque 320, a
multi-slice helical CT of the chest, abdomen, and pelvis was performed
during the venous phase of contrast enhancement. Additional oral
contrast was given using Gastrografin (diluted 1:35). Thin slice
reconstructions were obtained, along with secondary coronal and sagittal
reconstructions.
[Thorax]{.underline}:
Uniform presentation of the apical thoracic sections when included. No
evidence of subclavian lymphadenopathy. Uniform visualization of the pectoral
tissues. No evidence of mediastinal lymphadenopathy. The anterior segment
of the left upper lobe (series 205, image 88 of 389) shows a subpleural
ground-glass opaque solid consolidation. There is an enlarged lymph node
in the left hilar region measuring approximately 1.2 cm laterally. Otherwise,
there are no signs of suspicious intrapulmonary markings, no new inflammatory
infiltrates, no pneumothorax, no pericardial effusion. In the upper inner
quadrant of the right breast there is an oval mass, DD cystadenoma,
DD glandular cluster (measuring 1.2 cm).
[Abdomen/pelvis: ]{.underline}
Dominant display of the gastrocolic junction; absence of oral contrast
in this zone prevents more detailed analysis. Uniformly displayed
hepatic tissue with no signs of focal, density-varied lesions. Portal
and liver veins are well filled. Liver with minor auxiliary liver.
Adrenal nodes thin on both sides. Natural kidneys on both sides. Urinary
sac with placed transurethral tube and intravesical gas pockets.
Gallbladder typical. Paravertebral and within vertebral and in the
domain of the superior hepatic artery multiple pronounced lymph nodes,
these up to a maximum of 8 mm. Typical representation of the organs in
the pelvic region.
[Skeleton: ]{.underline}
Condition post dorsal reinforcement (T8-T9-L2-L3). After surgery,
epidermal air pockets and bloated tissue inflammation in the access path
zone. Signs of a resin in a pre-spinal vessel anterior to T8 and T9.
Known mixed osteoblastic/osteolytic bony metastasis of lumbar vertebra 4
and the cap plate of lumbar vertebra 2. State post-cutting of the
pedicle of T10. L5 also with slightly multiple solidified core
osteolytic defects.
[Evaluation:]{.underline}
- No sign of primary malignancy in the previously mentioned mixed
osteoblastic/osteolytic lesions in the vertebra (to be deemed
suspicious in coordination with the MR review of 12/20/2016).
- A magnified lymph node exists in the left hilar territory. In the
anterior left upper lobe, evidence of a solid cloudy consolidation,
i.e., of inflammatory origin, revisitation in 5-7 weeks recommended.
- Rounded consolidation in the upper inner quadrant of the right
breast, further breast examination advised.
**Functional Diagnostics**
Extended Respiratory Function (Diffusion) from 01/15/2017
[Evaluation]{.underline}
Patient cooperation: satisfactory. No detectable obstructive ventilation
issue. No pulmonary over-expansion after RV/TLC. No identified
restrictive ventilation impairment. Standard O2 diffusion ability. No
evidence of low oxygen levels, no blockage.
[Consultations / Therapy Reports]{.underline}
Psychological Support Consultation from 01/22/2017
[Current Situation/History:]{.underline}
The patient initially discussed \"night episodes\" in the calves, which
over time manifested during the day and were coupled with discomfort in
the cervical region. Prior, she had visited the Riverside Medical Center
multiple times before an MRI was executed. A \"mass in the neck\" was
identified. Since she suffers from fear of heights and fear of crowds,
the MRI could only be done under mild sedation. The phobias emerged
abruptly in 2011 with no apparent cause, leading to multiple hospital
visits. She is now in outpatient care. Additionally, she battles with
MS, with the most recent flare-up in 2012. She declined a procedure,
which was set for the MRI is day, because \"two sedatives in one day
felt excessive.\" She anticipates avoiding a repeated procedure.
Currently, however, she still experiences spasms in her right hand and a
numbing sensation in her fingers. She still encounters discomfort (NRS
5/10). She was previously informed that relief might be gradual, but she
is \"historically been restless\". Therefore, \"resting and inactivity\"
negatively impact her spirits and rest.
[Medical background:]{.underline}
Several in-patient and day clinic admissions since 2011.
Now, from 2015, continuous outpatient psychological counseling (CBT),
somatic therapy, particular sessions for driving anxieties. Also
undergoing outpatient psychiatric care (fluoxetine 90mg).
[Psychopathological Observations:]{.underline}
Patient appears well-groomed, responsive and clear-minded, talkative and
forthright. Aware of location, date, and identity. Adequate focus,
recall, and concentration. Mental organization is orderly. No evidence
of delusional beliefs or identity disturbances. No compulsions, mentions
fear of expansive spaces and fear of water. Emotional responsiveness
intact, heightened psychomotor activity. Mood swings between despondent
and irritable, lowered motivation. Diminished appetite, issues with
sleep initiation and maintenance. Firm and believable denial of
immediate suicidal thoughts, patient appears cooperative. No current
signs of self-harm or threat to others.
[Handling the Condition, Strengths:]{.underline}
Currently, her coping strategy seems to be proactive with some restless
elements. Ms. S. says she remains \"optimistic\" and is well-backed by
her communal links. Notably, she shares a close bond with her
80-year-old aunt. Her other social bonds primarily arise from her
association with a hockey enthusiasts club. Hockey has been a crucial
support for her from a young age.
[Evaluation Diagnoses:]{.underline}
Adjustment disorder: anxiety and depressive reaction mixed
Agoraphobia
Acrophobia
[Interventions, approaches:]{.underline}
An evaluative and assistive discussion was conducted. The patient has a
dependable therapeutic community for post-hospitalization. Additionally,
she was provided the contact of the psychological support outpatient
center. She mentioned finding the therapeutic conversation comforting,
prompting an arranged check-in the subsequent week. We also suggest
guidance in self-initiated physical activities to aid her recovery and
temper restlessness.
**NC: Consultation of 01/15/2017**
[Examination findings:]{.underline}
Patient alert, fully oriented. Articulate and spontaneous speech.
Cranial nerve evaluation normal. No evident sensorimotor abnormalities.
BDK with voiding challenges. Sphincter response diminished, but fecal
control maintained. KPPS at 85%. Wound site clean and non-irritated,
except for the lower central portion.
[Procedure:]{.underline}
Neurosurgical intervention not required; no reassessment of the lower
wound needed. Advise return if neurological symptoms intensify.
The patient, diagnosed with relapsing-remitting multiple sclerosis that
initially manifested aggressively, has been relapse-free on fingolimod
since 2009 and was generally well, barring a slight imbalance when
walking due to minor weakness in her left leg. She later experienced
numbness and weakness in her legs, reaching up to the hip, persisting
for several days and then faced challenges with urination and bowel
movements approximately 7 weeks prior. During a home examination, a
lesion was identified in the T10 which was surgically addressed by our
in-house neurosurgery team. Histology identified it as a DLBCL, leading
to a chemotherapy plan inclusive of Rituximab. Post-surgery, her
symptoms have subsided somewhat, but the patient still has BDK and
relies on a wheelchair.
On clinical neurological assessment, a mild paraparesis was noted in her
left leg, accompanied by heightened reflex response and sporadic left
foot spasms, which were intense but temporary.
To conclude, the new neurological manifestations are not a recurrence of
the formerly stable multiple sclerosis. As Rituximab is also an
effective third-phase drug for MS treatment and is essential,
discontinuing fingolimod (second phase) was discussed with the patient.
After a span of approximately 4-5 months following the last Rituximab
treatment, a radiological (cMRI) and clinical review is suggested. Based
on results, either resuming fingolimod or, if no adverse effects
present, potentially continuing Rituximab treatment is recommended (for
this, reach our neuroimmunology outpatient department). The primary
neurologist was unavailable for comments.
**Boards**
Oncology tumor board as of 01/22/2017
6 cycles of R-Pola-CHP
[Pathology]{.underline}
Pathology. Findings from 01/05/2017
[Clinical information/question:]{.underline}
Tumor cuff T10. dignity? Entity? Macroscopy:
1st lamina T10: fixed. some assembled 0.7 x 0.5 x 0.2 cm calcareous
tissue fragments. Complete embedding. Decalcify in EDTA. 2nd ligament:
Fix. some assembled 0.9 x 0.7 x 0.2 cm, coarse, partly also calcareous
tissue fragments. Complete embedding. Decalcify overnight in EDTA.
3\. epidural tumor: Numerous beige-colored tissue fragments, 3.8 x 2.8 x
0.6 cm. Embedding of exemplary sections after lamellation.
[Processing]{.underline}: 1 block, HE. Microscopy:
1\. and 2. (lamina T10 and ligament) are still being decalcified.
3rd epidural tumor: Paravertebral soft tissue with infiltrates of a
partly lymphoid, partly blastic neoplasia. The tumor cells are diffuse,
sometimes nodular in the tissue and have hyperchromatic nuclei with
coarse-grained chromatin and a narrow cytoplasmic border. There are also
blastic cells with enlarged nuclei, vesicular chromatin, and sometimes
prominent nucleoli. The stroma is loose and vacuolated. Clearly
pronounced crush artifacts.
Preliminary report of critical findings:
3\. epidural tumor: paravertebral soft tissue with infiltrates of
lymphoid and blastic cells compatible with hematologic neoplasia.
Additional immunohistochemical staining is being performed to further
characterize the tumor. In addition, material 1 (lamina T10) and
material 2 (ligament) are still undergoing decalcification. A follow-up
report will be provided.
Processing: 2 blocks, decalcification, HE, Giemsa, IHC: CD20, PAX5,
Bcl2, Bcl6, CD5, CD3, CD23, CD21, Kappa, Lambda, CD10, c-Myc, CyclinD1,
CD30, MIB1, EBV/EBER.
Molecular pathology: testing for B-cell clonal expansion and IgH/Bcl2
translocation.
[Microscopy]{.underline}:
1\. Ligament: Scarred connective tissue and fragmented bone tissue
without evidence of the tumor described in the preliminary findings
under 3.
2\. Lamina T10: Bone tissue without evidence of the tumor described in
the preliminary findings under 3.
3\. Epidural tumor: Immunohistochemically, blastic tumor cells show a
positive reaction after incubation with antibodies against CD20, PAX5
and BCL2. Partially positive reaction against Bcl-6 (\<20%). Some
isolated blastic cells staining positive for CD30. Lymphoid cells
positive for CD3 and CD5. Some residual germinal centers with positive
reaction to CD23 and CD21. Predominantly weak positive reaction of
blasts and lymphoid cells to CD10. Some solitary cells with positive
reaction to kappa, rather unspecific, flat reaction to lambda. No
overexpression of c-Myc or cyclinD1. No
No reaction to EBV/EBER. The Ki-67 proliferation index is 40%, related
to blastic tumor cells \> 90%.
Significantly limited evaluability of immunohistochemical staining due
to severe squeezing artifacts of the material.
[Molecular pathology report:]{.underline}
Examination for clonal B-cell expansion and t(14;18) translocation
Methodology:
DNA was isolated from the sent tissues and used in duplicate in specific
PCRs (B-cell clonality analysis with Biomed-2 primer sets: IGHG1 gene:
scaffold 2 and 3). The size distribution of the PCR products was further
analyzed by fragment analysis.
To detect a BCL2/IgH fusion corresponding to a t(14;18) translocation,
DNA was inserted into a specific PCR (according to Stetler-Stevenson et
al. Blood. 1998;72:1822-25).
Results:
Amplification of isolated DNA: good. B cell clonality analyses
IGHG1 fragment 2: polyclonal signal pattern.
IGHG1 frame 3: reproducible clonal signal at approximately 115/116 bp.
t(14;18) translocation: negative.
[Molecular pathology report:]{.underline}
Molecular pathologic evidence of clonal B-cell expansion. No evidence of
t(14;18) translocation in test material with normal control reactions.
Preliminary critical findings report:
1\. Lamina TH 10: tumor-free bone tissue.
2\. Ligament: Tumor-free, scarred connective tissue and fragmented bone
tissue.
3\. Epidural tumor: aggressive B non-Hodgkin\'s lymphoma.
Findings (continued)
Additional findings from 01/06/2017
Immunohistochemical processing: MUM1, ALK1, CD68, TdT. Microscopy:
3\. Immunohistochemically, blastic tumor cells are positive for MUM1.
Numerous CD68-positive macrophages. No reaction to ALK1 and TdT.
Critical findings report:
1\. Lamina T10: Tumor-free bone tissue. 2: Tumor free, scarred connective
tissue and fragmented bone tissue.
3\. epidural tumor: aggressive B-non-Hodgkin lymphoma, morphologically
and immunohistochemically most compatible with diffuse large B-cell
lymphoma (DLBCL, NOS) of germinal center type according to Hans
classifier (GCB).
**Path. Findings from 01/05/2017**
Clinical Findings
Clinical data:
Initial diagnosis of DLBCL with spinal involvement.
Puncture Site(s): 1
Collection date: 01/04/2017
Arrival at cytology lab: 01/04/2017, 8 PM. Material:
1 Liquid Material: 2 mL colorless, clear Processing:
MGG staining Microscopic:
ZTA:
Liquid precipitate Erythrocytes
(+) Lymphocytes (+) Granulocytes
Eosinophils Histiocytes Siderophages
\+ Monocytes
Others: Isolated evidence of fewer monocytes. No evidence of atypical
cells. Critical report of findings:
CSF sediment without evidence of inflammation or malignancy. Diagnostic
Grading:
Negative
Therapy and course
Mrs. S was admitted from the neurosurgical department for chemotherapy
(R-POLA-CHP) of suspected DLBCL with spinal/vertebral manifestations.
After exclusion of clinical and laboratory contraindications,
antineoplastic therapy was started on 01/08/2017. This was well
tolerated under the usual supportive measures. There were no acute
complications.
During her hospitalization, Ms. S reported numbness in both vascular
hemispheres. A neurosurgical and neurological presentation was made
without acute need for action. In consultation with the neurology
department, the existing therapy with fingolimod should be discontinued
due to the concomitant use of rituximab and the associated risk of PML.
If necessary, re-exposure to fingolimod may be considered after
completion of oncologic therapy.
On 01/07/2017, a port placement was performed by our vascular surgery
department without complications.
On 01/19/2017, a single administration of Pegfilgrastim 6 mg s.c. was
performed. With a latency of 10 days, G-CSF should not be repeated in
the meantime.
We are able to transfer Mrs. S to the Mountain Hospital Center
(Neurological Initial Therapy & Recovery) on 02/01/2017. We thank you
for accommodating the patient and are available for any additional
inquiries.
**Medications at Discharge**
**Aspirin (Aspirin®)** - 100mg, 1 tablet in the morning
**Atorvastatin - 40mg -** 1 tablet at bedtime
**Sertraline - 50mg** - 2 tablets in the morning
**Lorazepam (Tavor®)** - 1mg, as needed
**Fingolimod** - 0.5mg, 1 capsule at bedtime, Note: Take a break as
directed
**Hydromorphone hydrochloride** - 2mg (extended-release), 2 capsules in
the morning and 2 capsules at bedtime
**Melatonin -** 2mg (sustained-release), 1 tablet at bedtime
**Baclofen (Lioresal®) -** 10mg, 1 tablet three times a day
**Pregabalin -** 75mg, 1 capsule in the morning and 1 capsule at bedtime
**MoviCOL® (Macrogol, Sodium chloride, Potassium chloride) -** 1 packet
three times a day, mixed with water for oral intake
**Pantoprazole -** 40mg, 1 tablet in the morning
**Colecalciferol (Vitamin D3) -** 20000 I.U., 1 capsule on Monday and
Thursday
**Co-trimoxazole -** 960mg, 1 tablet on Monday, Wednesday, and Friday
**Valaciclovir -** 500mg, 1 tablet in the morning and 1 tablet at
bedtime
**Prednisolone -** 50mg, 2 tablets in the morning, Continue through
02/19/2017
**Enoxaparin sodium (Clexane®) -** 40mg (4000 I.U.), 1 injection at
bedtime, Note: Continue in case of immobility
**Dimenhydrinate (Vomex A®)** - 150mg (sustained-release), as needed for
nausea, up to 2 capsules daily.
**Procedure**
**Oncology board decision: 6 cycles of R-Pola-CHP.**
- Fingolimod pause, re-evaluation in 4-5 months.
- Continuation of therapy near residence in the clinic as of
02/28/2017
- Bi-Weekly laboratory tests (electrolytes, blood count, kidney and
liver function tests)
- In case of fever \>38.3 °C please report immediately to our
emergency room
- Immediate gynecological examination for nodular mass in the left
breast
Dates:
- From 03/01/2017 third cycle of R-Pola-CHP in the clinic. The patient
will be informed of the date by telephone.
If symptoms persist or exacerbate, we advocate for an urgent revisit.
Outside standard working hours, emergencies can also be addressed at the
emergency hub.
During discharge management, the patient was extensively educated and
assisted, and equipped with required appliances, medication scripts, and
absence from work notices.
All observations were thoroughly deliberated upon. Multiple alternate
therapy notions were considered before making a treatment proposition.
The opportunity for a second viewpoint and recommendation to our
facility was also emphasized.
**Lab values at discharge: **
**Metabolic Panel**
**Parameter** **Results** **Reference Range**
---------------------------------- ------------- ---------------------
Sodium 136 mEq/L 135 - 145 mEq/L
Potassium 3.9 mEq/L 3.5 - 5.0 mEq/L
Creatinine 1.2 mg/dL 0.7 - 1.3 mg/dL
BUN (Blood Urea Nitrogen) 19 mg/dL 7 - 18 mg/dL
Alkaline Phosphatase 138 U/L 40 - 129 U/L
Total Bilirubin 0.3 mg/dL \< 1.2 mg/dL
GGT (Gamma-Glutamyl Transferase) 82 U/L \< 66 U/L
ALT (Alanine Aminotransferase) 42 U/L 10 - 50 U/L
AST (Aspartate Aminotransferase) 34 U/L 10 - 50 U/L
LDH (Lactate Dehydrogenase) 366 U/L \< 244 U/L
Uric Acid 4.1 mg/dL 3 - 7 mg/dL
Calcium 9.0 mg/dL 8.8 - 10.6 mg/dL
**Kidney Function**
**Parameter** **Results** **Reference Range**
------------------------------- ------------- ---------------------
GFR (MDRD) \>60 mL/min \> 60 mL/min
GFR (CKD-EPI with Creatinine) 64 mL/min \> 90 mL/min
**Inflammatory Markers**
**Parameter** **Results** **Reference Range**
-------------------------- ------------- ---------------------
CRP (C-Reactive Protein) 2.5 mg/dL \< 0.5 mg/dL
**Coagulation Panel**
**Parameter** **Results** **Reference Range**
--------------- ------------- ---------------------
PT Percentage 103% 70 - 120%
INR 1.0 N/A
aPTT 25 sec 26 - 37 sec
**Complete Blood Count**
**Parameter** **Results** **Reference Range**
--------------- ---------------- ---------------------
WBC 12.71 x10\^9/L 4.0 - 9.0 x10\^9/L
RBC 2.9 x10\^12/L 4.5 - 6.0 x10\^12/L
Hemoglobin 8.1 g/dL 14 - 18 g/dL
Hematocrit 24.7% 40 - 48%
MCH 28 pg 27 - 32 pg
MCV 86 fL 82 - 92 fL
MCHC 32.8 g/dL 32 - 36 g/dL
Platelets 257 x10\^9/L 150 - 450 x10\^9/L
**Differential**
**Parameter** **Results** **Reference Range**
--------------- ------------- ---------------------
Neutrophils 77% 40 - 70%
Lymphocytes 4% 25 - 40%
Monocytes 18% 4 - 10%
Eosinophils 0% 2 - 4%
Basophils 0% 0 - 1%
### Patient Report 1
**Dear colleague, **
I am writing to provide a follow-up report on our mutual patient, Mrs.
Anna Sample, born on January 1st, 1970, post her recent visit to our
clinic on October 9th, 2017.
Upon assessment, Mrs. Sample reported experiencing a moderate
improvement in symptoms since the initiation of the R-Pola-CHP regimen.
The discomfort in her dorsal calf and thoracic spine has notably
reduced, and her arm strength has seen gradual improvement, though she
occasionally still encounters difficulty in grasping objects.
She has been undergoing physiotherapy to aid in the recovery of her arm
strength.
**Physical Examination:** No palpable lymphadenopathy. Her neurological
examination was stable with no new deficits.
**Laboratory Findings:** Most recent blood counts and biochemistry
panels showed a trend towards normalization, with liver enzymes within
the reference range.
**Imaging:**
-Ultrasound of the abdomen was conducted.
-A follow-up MRI conducted showed a reduction in the size of the
previously noted metastatic masses. There\'s a decreased impingement on
the myelon at the levels of T10-L1. The lesions in L2, L3, and L4 also
showed signs of regression.
-PET scan was performed: Favourable response. Increased FDG avidity in
the liver: Liver MRI recommended.
-Liver MRI: No pathology of the liver.
**Senological Examination:** The nodule-like deposit in the right breast
was found to be benign.
**Medication on admission:** Aspirin (Aspirin®) - 100mg, 1 tablet in the
morning Atorvastatin - 40mg - 1 tablet at bedtime Sertraline - 50mg - 2
tablets in the morning Lorazepam (Tavor®) - 1mg, as needed Fingolimod -
0.5mg, 1 capsule at bedtime, Note: Take a break as directed
Hydromorphone hydrochloride - 2mg (extended-release), 2 capsules in the
morning and 2 capsules at bedtime Melatonin - 2mg (sustained-release), 1
tablet at bedtime Baclofen (Lioresal®) - 10mg, 1 tablet three times a
day Pregabalin - 75mg, 1 capsule in the morning and 1 capsule at bedtime
MoviCOL® (Macrogol, Sodium chloride, Potassium chloride) - 1 packet
three times a day, mixed with water for oral intake Pantoprazole - 40mg,
1 tablet in the morning Colecalciferol (Vitamin D3) - 20000 I.U., 1
capsule on Monday and Thursday Co-trimoxazole - 960mg, 1 tablet on
Monday, Wednesday, and Friday Valaciclovir - 500mg, 1 tablet in the
morning and 1 tablet at bedtime Prednisolone - 50mg, 2 tablets in the
morning, Continue through 02/19/2017 Enoxaparin sodium (Clexane®) - 40mg
(4000 I.U.), 1 injection at bedtime, Note: Continue in case of
immobility Dimenhydrinate (Vomex A®) - 150mg (sustained-release), as
needed for nausea, up to 2 capsules daily.
**Physician\'s report for ultrasound on 10/05/2017:**
Liver: The liver is large with 18.1 cm in the MCL, 18.5 cm in the CCD
and 20.2 cm in the AL. The internal structure is not compacted. Focal
changes are not seen. Orthograde flow in the portal vein (vmax 16 cm/s).
Gallbladder: the gallbladder is 9.0 x 2.9 cm, the lumen is free of
stones.
Biliary tract: The intra- and extrahepatic bile ducts are not
obstructed, DHC 5 mm, DC 3 mm.
Pancreas: The pancreas is approximately 3.2/1.5/3.0 cm in size, the
internal structure is moderately echo-rich.
Spleen: The spleen is 28.0 x 9.6 cm, the parenchyma is homogeneous.
Kidneys: The right kidney is 9.8/2.0 cm, the pelvis is not congested.
The left kidney is 12.4/1.2 cm, the pelvis is not congested. Vessels
retroperitoneal: the aorta is normal in width in the partially visible
area.
Stomach/intestine: The gastric corpus wall is up to 14 mm thick. No
evidence of free fluid in the abdominal cavity.
Bladder/genitals: The prostate is orientationally about 3.8 x 4.8 x 3.1
cm, the urinary bladder is moderately full.
**MR Spine plain + post-contrast from 10/06/2017**
**Study:** Magnetic Resonance Imaging (MRI) of the thoracolumbar spine
**Clinical Information:** Follow-up MRI post treatment for previously
noted metastatic masses.
**Technique:** Standard T1-weighted, T2-weighted, and post-contrast
enhanced sequences of the thoracolumbar spine were obtained in sagittal
and axial planes.
**Findings:** There is a reduction in the size of the previously noted
metastatic masses when compared to prior MRI studies. A reduced mass
effect is observed at the levels of T10-L1. Notably, there is decreased
impingement on the myelon at these levels. This indicates a significant
improvement, suggesting a positive response to the recent therapy. The
lesion noted in the previous study at the level of L2 has shown signs of
regression in both size and intensity. Similar regression is noted for
the lesion at the L3 level. The lesion at the L4 level has also
decreased in size as compared to previous imaging. The intervertebral
discs show preserved hydration. No significant disc protrusions or
herniations are observed. The vertebral bodies do not show any
significant collapse or deformity. Bone marrow signal is otherwise
normal, apart from the aforementioned lesions. The spinal canal
maintains a normal caliber throughout, and there is no significant canal
stenosis. The conus medullaris and cauda equina nerve roots appear
unremarkable without evidence of displacement or compression.
**Impression:** Reduction in the size of previously noted metastatic
masses, indicating a positive therapeutic response. Decreased
impingement on the myelon at the levels of T10-L1, suggesting
significant regression of the previously observed mass effect.
Regression of lesions at L2, L3, and L4 levels, further indicating the
positive response to treatment.
**Positron Emission Tomography (PET)/CT from 10/09/2017:**
**Indication:** Follow-up evaluation of Diffuse large B-cell lymphoma of
germinal center type diagnosed in 01/2017.
**Technique:** Whole-body FDG-PET/CT was performed from the base of the
skull to the mid-thighs.
**Findings:** Liver: There is increased FDG uptake in the liver,
predominantly in the anterolateral segment. The size of the liver is
consistent with the previous ultrasound report, measuring 18.1 cm in the
MCL, 18.5 cm in the CCD, and 20.2 cm in the AL. The SUV max is 5.5.
Lymph Nodes: There is no pathological FDG uptake in the previously noted
left axillary lymph node, suggesting a therapeutic response. Lungs:
Previously noted deposit in the ventral left upper lobe now demonstrates
reduced FDG avidity. No other FDG-avid nodules or masses. Bone: There\'s
no FDG uptake in the spine, including the previously described
metastatic lesion, indicating a positive response to treatment.
**Impression:** Overall, the findings demonstrate a marked metabolic
improvement in the sites of lymphoma previously noted, particularly in
the left axillary lymph node and the vertebral bone lesions. The liver,
however, presents with increased FDG avidity, especially in the
anterolateral segment. This uptake might represent active lymphomatous
involvement or could be due to an inflammatory process. Given the
differential, and to ascertain the etiology, further diagnostic
evaluation, such as a liver MRI or biopsy, is recommended.
**Liver MRI from 10/11/2017:**
**Clinical Indication:** Evaluation of increased FDG uptake in the liver
as noted on the recent PET scan. Concern for active lymphomatous
involvement or an inflammatory process.
**Technique:** MRI of the liver was performed using a 3T scanner.
Sequences included T1-weighted (in-phase and out-of-phase), T2-weighted,
diffusion-weighted imaging (DWI), and post-contrast dynamic imaging
after the administration of gadolinium-based contrast agent.
**Detailed Findings:** The liver demonstrates enlargement with
measurements consistent with the recent ultrasound: 18.1 cm in the
mid-clavicular line (MCL), 18.5 cm in the maximum cranial-caudal
diameter (CCD), and 20.2 cm along the anterior line (AL).
The liver parenchyma is mostly homogenous. However, there is a region in
the anterolateral segment demonstrating T2 hyperintensity and
hypointensity on T1-weighted images. The aforementioned region in the
anterolateral segment demonstrates restricted diffusion, suggestive of
increased cellular density. After gadolinium administration, there is
peripheral enhancement of the lesion in the arterial phase, followed by
progressive central filling in portal venous and delayed phases. This
pattern is suggestive of a focal nodular hyperplasia (FNH) or atypical
hemangioma. The intrahepatic and extrahepatic bile ducts are not
dilated. No evidence of any obstructing lesion. The hepatic arteries,
portal vein, and hepatic veins appear patent with no evidence of
thrombosis or stenosis. The gallbladder, pancreas, spleen, and adjacent
segments of the bowel appear normal. No lymphadenopathy is noted in the
porta hepatis or celiac axis.
**Impression:** Enlarged liver with a suspicious lesion in the
anterolateral segment demonstrating characteristics that might be
consistent with focal nodular hyperplasia or atypical hemangioma. No
indication of lymphomatous involvement of the liver.
**Discussion: **
Given her positive response to the treatment so far, we intend to
continue with the current regimen, with careful monitoring of her side
effects and symptomatology.
We deeply appreciate your continued involvement in Mrs. Sample\'s
healthcare journey. Collaborative care is paramount, especially in cases
as complex as hers. Should you have any recommendations, insights, or if
you require additional information, please do not hesitate to reach out.
**Medication at discharge: **
Aspirin 100mg: Take 1 tablet in the morning; Atorvastatin 40mg: Take 1
tablet at bedtime; Sertraline 50mg: Take 2 tablets in the morning;
Lorazepam 1mg: Take as needed; Melatonin (sustained-release) 2mg: Take 1
tablet at bedtime; Fingolimod 0.5mg: Take 1 capsule at bedtime/take a
break as directed; Hydromorphone hydrochloride (extended-release) 2mg:
Take 2 capsules in the morning and 2 capsules at bedtime; Pregabalin
75mg: Take 1 capsule in the morning and 1 capsule at bedtime; Baclofen
10mg: Take 1 tablet three times a day; MoviCOL®: Mix 1 packet with water
and take orally three times a day; Pantoprazole 40mg: Take 1 tablet in
the morning; Colecalciferol (Vitamin D3) 20000 I.U.: Take 1 capsule on
Monday and Thursday; Dimenhydrinate (sustained-release) 150mg: Take as
needed for nausea, up to 2 capsules daily.
**Metabolic Panel**
**Parameter** **Results** **Reference Range**
---------------------------------- ------------- ---------------------
Sodium 138 mEq/L 135 - 145 mEq/L
Potassium 4.1 mEq/L 3.5 - 5.0 mEq/L
Creatinine 1.1 mg/dL 0.7 - 1.3 mg/dL
BUN (Blood Urea Nitrogen) 17 mg/dL 7 - 18 mg/dL
Alkaline Phosphatase 124 U/L 40 - 129 U/L
Total Bilirubin 0.4 mg/dL \< 1.2 mg/dL
GGT (Gamma-Glutamyl Transferase) 75 U/L \< 66 U/L
ALT (Alanine Aminotransferase) 39 U/L 10 - 50 U/L
AST (Aspartate Aminotransferase) 36 U/L 10 - 50 U/L
LDH (Lactate Dehydrogenase) 342 U/L \< 244 U/L
Uric Acid 3.8 mg/dL 3 - 7 mg/dL
Calcium 9.12 mg/dL 8.8 - 10.6 mg/dL
**Kidney Function**
**Parameter** **Results** **Reference Range**
------------------------------- ------------- ---------------------
GFR (MDRD) \>62 mL/min \> 60 mL/min
GFR (CKD-EPI with Creatinine) 67 mL/min \> 90 mL/min
**Inflammatory Markers**
**Parameter** **Results** **Reference Range**
-------------------------- ------------- ---------------------
CRP (C-Reactive Protein) 1.8 mg/dL \< 0.5 mg/dL
**Coagulation Panel**
**Parameter** **Results** **Reference Range**
--------------- ------------- ---------------------
PT Percentage 105% 70 - 120%
INR 0.98 N/A
aPTT 28 sec 26 - 37 sec
**Complete Blood Count**
**Parameter** **Results** **Reference Range**
--------------- --------------- ---------------------
WBC 11.9 x10\^9/L 4.0 - 9.0 x10\^9/L
RBC 3.1 x10\^12/L 4.5 - 6.0 x10\^12/L
Hemoglobin 8.4 g/dL 14 - 18 g/dL
Hematocrit 26% 40 - 48%
MCH 27.8 pg 27 - 32 pg
MCV 84 fL 82 - 92 fL
MCHC 33 g/dL 32 - 36 g/dL
Platelets 263 x10\^9/L 150 - 450 x10\^9/L
**Differential**
**Parameter** **Results** **Reference Range**
--------------- ------------- ---------------------
Neutrophils 73% 40 - 70%
Lymphocytes 7% 25 - 40%
Monocytes 16% 4 - 10%
Eosinophils 1% 2 - 4%
Basophils 0.5% 0 - 1% | Mild paraparesis in the left leg |
Is there any explanation why some choice of language pair is better than the other? | ### Introduction
The spread of influenza is a major health concern. Without appropriate preventative measures, this can escalate to an epidemic, causing high levels of mortality. A potential route to early detection is to analyse statements on social media platforms to identify individuals who have reported experiencing symptoms of the illness. These numbers can be used as a proxy to monitor the spread of the virus. Since disease does not respect cultural borders and may spread between populations speaking different languages, we would like to build models for several languages without going through the difficult, expensive and time-consuming process of generating task-specific labelled data for each language. In this paper we explore ways of taking data and models generated in one language and transferring to other languages for which there is little or no data. ### Related Work
Previously, authors have created multilingual models which should allow transfer between languages by aligning models BIBREF0 or embedding spaces BIBREF1, BIBREF2. An alternative is translation of a high-resource language into the target low-resource language; for instance, BIBREF3 combined translation with subsequent selective correction by active learning of uncertain words and phrases believed to describe entities, to create a labelled dataset for named entity recognition. ### MedWeb Dataset
We use the MedWeb (“Medical Natural Language Processing for Web Document”) dataset BIBREF4 that was provided as part of a subtask at the NTCIR-13 Conference BIBREF5. The data is summarised in Table TABREF1. There are a total of 2,560 pseudo-tweets in three different languages: Japanese (ja), English (en) and Chinese (zh). These were created in Japanese and then manually translated into English and Chinese (see Figure FIGREF2). Each pseudo-tweet is labelled with a subset of the following 8 labels: influenza, diarrhoea/stomach ache, hay fever, cough/sore throat, headache, fever, runny nose, and cold. A positive label is assigned if the author (or someone they live with) has the symptom in question. As such it is more than a named entity recognition task, as can be seen in pseudo-tweet #3 in Figure FIGREF2 where the term “flu” is mentioned but the label is negative. ### Methods ::: Bidirectional Encoder Representations from Transformers (BERT):
The BERT model BIBREF6 base version is a 12-layer Transformer model trained on two self-supervised tasks using a large corpus of text. In the first (denoising autoencoding) task, the model must map input sentences with some words replaced with a special “MASK” token back to the original unmasked sentences. In the second (binary classification) task, the model is given two sentences and must predict whether or not the second sentence immediately follows the first in the corpus. The output of the final Transformer layer is passed through a logistic output layer for classification. We have used the original (English) BERT-base, trained on Wikipedia and books corpus BIBREF7, and a Japanese BERT (jBERT) BIBREF8 trained on Japanese Wikipedia. The original BERT model and jBERT use a standard sentence piece tokeniser with roughly 30,000 tokens. ### Methods ::: Multilingual BERT:
Multilingual BERT (mBERT) is a BERT model simultaneously trained on Wikipedia in 100 different languages. It makes use of a shared sentence piece tokeniser with roughly 100,000 tokens trained on the same data. This model provides state-of-the-art zero-shot transfer results on natural language inference and part-of-speech tagging tasks BIBREF9. ### Methods ::: Translation:
We use two publicly available machine translation systems to provide two possible translations for each original sentence: Google's neural translation system BIBREF10 via Google Cloud, and Amazon Translate. We experiment using the translations singly and together. ### Methods ::: Training procedure:
Models are trained for 20 epochs, using the Adam optimiser BIBREF11 and a cyclical learning rate BIBREF12 varied linearly between $5 \times 10^{-6}$ and $3 \times 10^{-5}$. ### Experiments
Using the multilingual BERT model, we run three experiments as described below. The “exact match” metric from the original MedWeb challenge is reported, which means that all labels must be predicted correctly for a given pseudo-tweet to be considered correct; macro-averaged F1 is also reported. Each experiment is run 5 times (with different random seeds) and the mean performance is shown in Table TABREF11. Our experiments are focused around using Japanese as the low-resource target language, with English and Chinese as the more readily available source languages. ### Experiments ::: Baselines
To establish a target for our transfer techniques we train and test models on a single language, i.e. English to English, Japanese to Japanese, and Chinese to Chinese. For English we use the uncased base-BERT, for Japanese we use jBERT, and for Chinese we use mBERT (since there is no Chinese-specific model available in the public domain). This last choice seems reasonable since mBERT performed similarly to the single-language models when trained and tested on the same language. For comparison, we show the results of BIBREF13 who created the most successful model for the MedWeb challenge. Their final system was an ensemble of 120 trained models, using two architectures: a hierarchical attention network and a convolutional neural network. They exploited the fact that parallel data is available in three languages by ensuring consistency between outputs of the models in each language, giving a final exact match score of 0.880. However, for the purpose of demonstrating language transfer we report their highest single-model scores to show that our single-language models are competitive with the released results. We also show results for a majority class classifier (predicting all negative labels, see Table TABREF1) and a random classifier that uses the label frequencies from the training set to randomly predict labels. ### Experiments ::: Zero-shot transfer with multilingual pre-training
Our first experiment investigates the zero-shot transfer ability of multilingual BERT. If mBERT has learned a shared embedding space for all languages, we would expect that if the model is fine-tuned on the English training dataset, then it should be applicable also to the Japanese dataset. To test this we have run this with both the English and Chinese training data, results are shown in Table TABREF11. We ran additional experiments where we froze layers within BERT, but observed no improvement. The results indicate poor transfer, especially between English and Japanese. To investigate why the model does not perform well, we visualise the output vectors of mBERT using t-SNE BIBREF14 in Figure FIGREF14. We can see that the language representations occupy separate parts of the representation space, with only small amounts of overlap. Further, no clear correlation can be observed between sentence pairs. The better transfer between Chinese and Japanese likely reflects the fact that these languages share tokens; one of the Japanese alphabets (the Kanji logographic alphabet) consists of Chinese characters. There is 21% vocabulary overlap for the training data and 19% for the test data, whereas there is no token overlap between English and Japanese. Our finding is consistent with previous claims that token overlap impacts mBERT's transfer capability BIBREF9. ### Experiments ::: Training on machine translated data
Our second experiment investigates the use of machine translated data for training a model. We train on the machine translated source data and test on the target test set. Results are shown in Table TABREF11. Augmenting the data by using two sets of translations rather than one proves beneficial. In the end, the difference between training on real Japanese and training on translations from English is around 9% while training on translations from Chinese is around 4%. ### Experiments ::: Mixing translated data with original data
Whilst the results for translated data are promising, we would like to bridge the gap to the performance of the original target data. Our premise is that we start with a fixed-size dataset in the source language, and we have a limited annotation budget to manually translate a proportion of this data into the target language. For this experiment we mix all the translated data with different portions of original Japanese data, varying the amount between 1% and 100%. The results of these experiments are shown in Figure FIGREF17. Using the translated data with just 10% of the original Japanese data, we close the gap by half, with 50% we match the single-language model, and with 100% appear to even achieve a small improvement (for English), likely through the data augmentation provided by the translations. ### Discussion and Conclusions
Zero-shot transfer using multilingual BERT performs poorly when transferring to Japanese on the MedWeb data. However, training on machine translations gives promising performance, and this performance can be increased by adding small amounts of original target data. On inspection, the drop in performance between translated and original Japanese was often a result of translations that were reasonable but not consistent with the labels. For example, when translating the first example in Figure FIGREF2, both machine translations map “UTF8min風邪”, which means cold (the illness), into “UTF8min寒さ”, which means cold (low temperature). Another example is where the Japanese pseudo-tweet “UTF8min花粉症の時期はすごい疲れる。” was provided alongside an English pseudo-tweet “Allergy season is so exhausting.”. Here, the Japanese word for hay fever “UTF8min花粉症。” has been manually mapped to the less specific word “allergies” in English; the machine translation maps back to Japanese using the word for “allergies” i.e. “UTF8minアレルギー” in the katakana alphabet (katakana is used to express words derived from foreign languages), since there is no kanji character for the concept of allergies. In future work, it would be interesting to understand how to detect such ambiguities in order to best deploy our annotation budget. Table 1: MedWeb dataset overview statistics. Table 2: Overall results, given as mean (standard deviation) of 5 runs, for different training/test data pairs. The leading results on the original challenge are shown as baselines for benchmarking purposes. EN - English, JA - Japanese, ZH - Chinese, TJA - Translated Japanese. Figure 2: Max-pooled output of mBERT final layer (before fine tuning), reduced using principal component analysis (to reduce from 768 to 50 dimensions) followed by t-SNE (to project into 2 dimensions). 20 sentence triplets are linked to give an idea of the mapping between languages. Figure 3: Exact match accuracy when training on different proportions of the original Japanese training set, with or without either the original English data or the translated data. The pink and orange dashed lines show the accuracy of the full set of translated Japanese data (from English and Chinese respectively) and the blue dashed line shows the accuracy of the full original Japanese data. | translations that were reasonable but not consistent with the labels |
What was the last animal left on Earth after the mysterious plague began to spread?
A. Household pets
B. Rats
C. Humans
D. Locusts
| "Phone Me in Central Park" By JAMES McCONNELL There should be an epitaph for every man, big or little, but a really grand and special one for Loner Charlie. [Transcriber's Note: This etext was produced from Planet Stories Fall 1954. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed.] Charles turned over on his side to look at her. She lay quietly in the other bed, the most beautiful woman he had ever seen. She was blonde to perfection, exquisitely shaped, and the rich promise of her body was exposed to his view. "Why?" he thought as he looked at her. "Why did it have to happen like this?" The whole thing was still like a dream to him, and as yet he couldn't decide whether it was a good or a bad dream. A year ago she had been unattainable, a face to conjure with in erotic dreams, far beyond his ken. A year ago she had been a public idol, the most popular actress of the day. And he had been a nobody, full of a nobody's idle hopes and schemes. And now he was lying in the bed next to hers in her swank Manhattan apartment in the most exclusive hotel in town. The unrealness of the situation overwhelmed him. His mind was a picture of confused thoughts. Meanings and answers to his questions slithered out of his reach. "God," he said. It was not an exclamation, nor yet an expletive. It was a mere statement of fact. A thought teased at him. Charles looked at the woman again and decided that she still looked beautiful in spite of the harshness of the room's lighting. He touched buttons by the edge of the bed and the illumination quieted to a soft glow, wrapping her in a radiant halo. Charles smiled wanly and got up. He stood by the bed looking at her. "I could have fallen in love with you once. A year ago, perhaps, or longer. But not now. Not now." He turned away and walked to the window. "Now the world is dead. The whole world is dead." New York lay quietly below him. It was the hour of indecision when day has not quite made up its mind to leave and night has not yet attacked in force. The streetlights were already on, making geometric patterns through the dusk of Central Park. Some of the billboards were shining, their relays activated by darkness-sensitized solenoids. A reddish-orange pallor hung from the sky. It had been very pleasant that afternoon. She had given of herself freely, warmly, and Charles had accepted. But then he had known that she would. It was not him, it was the circumstances. Under the circumstances, she would have given herself to any man— "Why did it have to be her—or me? Why should it have to happen to anybody! Why!" She would have given herself to any man— His thoughts beat a rapid crescendo, activating emotions, stimulating sensations of angry rage. He wanted to cry, to weep angry tears of protest. To any man, WHO HAPPENED TO BE THE LAST MAN ON EARTH! Charles picked up a heavy book end off the table and crashed it through the thick pane of window glass. A gust of wind from the outside breezed through the shattered opening, attacking his olfactory patch with the retching smell of decaying flesh. Charles ignored it. Even smells had lost their customary meanings. He felt the rage build up inside again, tearing at his viscera. His stomach clenched up like an angry fist. "But I don't want to be the last man alive!" he shouted. "I don't know what to do! I don't know where to go, how to act! I just don't know—" A paroxysm of sobbing shook his body. Trembling, he dropped to his knees, his head against the cold firmness of the sill, his hands clutched tightly around the jagged edges of the window pane. In spite of the sharp pain that raced through his system, in spite of the bright, warm, red stream that trickled down his face, he knelt by the window for several minutes. " Maybe I'm not the last! " The thought struck him with suddenness, promisingly, edged with swelling comfort to fill his emptiness. Charles got up slowly, noticing for the first time that his fingers were badly cut. He wrapped a handkerchief around them and forgot them. He had to know—he had to find out. As he turned to leave, he noticed again the woman lying in radiant state upon the bed. He walked to her side and leaned over, kissing her gently on the forehead. As he straightened up, his leg caught against her arm, pushing it slightly. The woman's arm slipped from its position and dangled from the edge of the bed like a crazy pendulum. Charles picked it up and folded it across her now cold breasts. He started to pull the sheet over her nude form, then stopped, smiling at his conventionality. After all, it didn't make any difference now. The phonograph was near the door. On sudden impulse he switched it on, turned the volume up full, and in grim jest left it playing Rachmaninoff's Isle of the Dead on full automatic. The music haunted him down the hall to the elevator that he had to run himself. The lobby was littered with debris, human and otherwise. Charles ignored it. The street that led towards the Bureau of Vital Statistics was a mess of desolate carnage. Charles overlooked it. Shop fronts smashed, stores looted, gyro-cars wrecked, proud buildings defaced. "That was it," he said to himself. "Pride. We called this the 'Proud Era.' Everything was better and bigger and nicer to have. Buildings were taller, men were healthier, most of the problems of humanity seemed licked, or nearly so. It was a time of free power, each small unit of population, each section of town operating on perpetual, ever-lasting, automatic atomic piles. "We were free. We seemed, almost, to have accomplished something. The world was running well. No wonder we called it the 'Proud Era.' Life was fun, just a bowl of cherries, until...." Two years ago the animals had started dying. Strangely enough the rats had gone first, to anybody's notice. Sales of poison dropped, scientific laboratories chained to a perpetual rodent-cycle began to complain bitterly. Then the lovers who hunted out and haunted the lonely lanes through the countryside began to remark that the locusts were late that year. The Southern states joyously reported that mosquito control was working to an unprecedented degree. The largest cotton crop ever was forecast and rumors from Mexico had it that no one had died from scorpion bite in several weeks. A month later the meat animals, the birds and the household pets began dropping as rapidly as the flies which had dropped earlier. Congress was called into special session, as were all of the national governments around the world. The U.N. met at emergency sessions to cope with the situation. The president of the world-wide Society for the Prevention of Cruelty to Animals committed suicide. Within a year it was obvious to everyone that man was the only animal left on earth. The panic which had begun with the death of the animals was quieted somewhat by the fact that humans seemed immune to the pandemic. But the lakes full of dead fish caused a great stink and residents along the coasts began to move inland. Sales of perfumes and deodorants soared. Then just one year ago, the first human became infected with the strange malady. Within six months, half of the world's population was gone. Less than a month ago no more than a few thousand people remained in New York. And now.... "I've got to find out," Charles told himself. He meant it, of course, but in a sense he was afraid—afraid that his trip to the Bureau might give him an answer he didn't dare listen to. "But I've got to try." He walked on down the bloody street. Before the plague the Bureau of Vital Statistics had been one of man's crowning achievements. Housed as it was in a huge metallic globe of a building, it contained computers which kept exact account of every human on earth. Compulsory registration and the classification of each individual by means of the discrete patterns of his brain waves had accomplished for man what no ordinary census could have. The machine knew who was alive, who was dead, and where everybody was. Once a year the Bureau issued The Index, an exact accounting of Earth's four billion inhabitants. Four billion names and addresses, compressed into microprint, a tremendous achievement even for the "Proud Era." In all of his life, Charles had never once glanced at The Index. The average person had little necessity to do so since the Bureau information service would answer questions free of charge at any time. Reaching the gigantic building, Charles pushed aside the body of a young man and walked into the main foyer. Passing behind once-guarded doors, he entered the giant computer room and paused in admiration. Only once, before the plague, had he seen the interior of this room. But he still remembered it and he still recalled the powerful emotional experience it had been those many years ago. All children had to have a brain-wave recording made by the Bureau during the first month of their life. And again at the age of 10 each child returned to the Bureau for a recheck. It was for this latter recording that Charles had come to the Bureau some twenty-two years before and a friendly guard had let him peep briefly into the computer room. The impression of intense activity, of organized confusion, of mechanical wonder had remained with him the rest of his life. "So different now," he thought, surveying the room. "Now it's empty, so empty." The machine seemed to reflect the stillness, the very deadness of the world. The silence became unbearable. Charles walked to the master control panel. With newly acquired dexterity he switched the computer screens on and watched them glow to life. All around the world sensitive receiving stations pulsed to activity, sending out searching fingers, hunting for elusive patterns of neutral energy, mapping and tabulating the results. The main computer screen dominated one wall of the room. Other smaller screens clustered around it. On these screens could be graphed the population of any and every part of the globe. An illuminated counter immediately above it would give the numerical strength of the area being sampled while the screen would show population density by individual pinpoints of light that merged to form brightness patterns. "I'll try New York first," he said to himself, knowing that he was a coward, afraid to check the whole world from the start. "I'll start with New York and work up." Charles activated the switches that would flash a schematic map of New York on the screen. "There's bound to be somebody else left here. After all, there were at least twenty of us just a couple of days ago." And one of them, a beautiful woman, had invited him up to her apartment, not because she liked him, but because.... The main screen focused itself, the patterns shifting into a recognizable perceptual image. "Why, it was just yesterday (or was it the day before?) that ten of us, at least, met here to check the figures. There were lots of us alive then." Including the blond young woman who had died just this afternoon.... Charles stopped talking and forced his eyes upwards. Peripheral vision caught first the vague outlines of the lower part of the map. His eyes continued to move, slowly, reluctantly. They caught the over-all relief of Greater New York City—and then concentrated on the single, shining dot at the very heart of the map—and he understood. His eyes stabbed quickly for the counter above the screen. One. He gasped. The counter read one . Charles was by himself, the last person alive in all of New York City. He began to tremble violently. The silence of the room began to press quickly in on him. His frantic fingers searched for the computer controls. New York State. One. The entire United States. One. The western hemisphere, including islands. (Was that a point of light in Brazil? No. Just a ghost image). One. The Pacific area, Asia, Australia, Asia Minor, Russia and the Near East, Africa and then Europe. England! There was a light in England! Someone else still lived! The counter clicked forward. Two! His trembling stopped. He breathed again. "Of course. London was at least as populous as New York City before the plague. It's only logical that—" He stopped. For even as he spoke, the light winked out! The counter clicked again. One. Alone. Alone! Charles screamed. The bottom dropped out from under him! Why? Such a simple question, but in those three letters lay the essence of human nature. Why. The drive of curiosity. Stronger, in a way, than the so-called "basic" drives: hunger, thirst, sex, shelter, warmth, companionship, elimination. Certainly more decisive in the history of the race. Man began to think, to differentiate himself from the other animals, when he first asked the question: "Why?" But thinking about "why" didn't answer the question itself, Charles thought. He looked around him. He was sitting on a bench in Central Park, alone except for a few stray corpses. But the park was fairly free of bodies. "You've got about ten minutes warning," he said to himself. "I guess that most people wanted to die inside of something—inside of anything. Not out in the unprotected open." The silence was like a weight hanging around his neck. Not an insect noise, not the chirp of a bird, not the sound of a car nor the scream of a plane. Not even a breeze to whisper among the leaves, he thought. Civilization equals life equals noise. Silence equals.... Why. His mind kept returning to the question. Of all the people on earth, me. The last. Why me? Average, that's what he was. Height: 5'11". Weight: 165. Age: 32. Status: Married, once upon a time. The Norm, with no significant departures, all down the line. Church member, but not a good one. Could that be it? Could the most normal be the most perfect? Had he led the best of all possible lives? Was that it? Had God, in His infinite wisdom and mercy, spared his life, saved him, singled him out because he was most nearly a saint, most nearly Christ-like, most nearly.... Lies—His mind snapped back to reality. He half smiled. Saint? Christ? The Second Coming? He was no saint. Charles sighed. What about—? Chance. That was it! The laws of probability, the bell-shaped curve, normal distribution, rectilinear regression. More people per square foot in New York than elsewhere. The first person who died was from New York, so the last person who gave way to the disease should come from here too. Spin the wheel; throw the dice; toss the coin. So simple to explain by the laws of chance. No need for any underlying assumptions about good and evil, no need for teleological arguments concerning cause and effect. Simply explain it by chance. Somebody had to be the last to go and that was— "No," Charles said, standing up in the quiet of the spring evening. "No, chance won't do it. No man can reckon with chance. The mind rejects such things. There must be something beyond mere accident. There must be!" He sighed slowly. "So now I'm a hermit, whether or not I like it," he said in derision to the gravel path as he walked along it. "A hermit in the midst of a city of millions of—No, I forgot. There aren't any more people, are there?" It was hard to realize, even now. "A hermit, alone—and I haven't even got a cave...." Charles stopped walking suddenly. No cave, he thought. No place to sleep out the long one, no place to rest while time came to change things around and make them for the better. No place to hide. And suddenly it was the most important thing in life to him to find his "cave." It took him almost an hour to find the proper tools, and better than two hours more of hard, nighttime work to get the hole dug to his satisfaction. It took almost three hours to find the right sort of casket, durable but not too heavy for one man to handle. He carted it out to a grassy plot close to the center of the park where the grave was. He let the coffin down slowly into the depression, then piled up loose dirt on the sloping sides of the hole so that the rain would wash it down over him. "I can't very well bury myself," he said. "I guess it will rain after I'm gone." He looked carefully down at the metallic container. Wait now. There was something wrong, something missing. It was—oh, yes, he caught it. It was the stone. There wasn't any stone to go at the head of the grave. "I'll have to fix that." A sheet of metal, bent double, served for the monument proper. A nearby tool shed yielded up a can of paint and a brush. By the glow of one of the streetlights Charles worked out the inscription. "It ought to be something impressive," he thought out loud. "Something fitting the occasion." What did one say on these situations? There was so little chance to practice up for things like this. But it ought to be good, it ought to be proper. "'In this now hallowed corner of the planet Earth—' No. That sounds too ... too...." Make it simple, he thought. And he finally wrote: HERE LIES THE BODY OF THE LAST MAN ON EARTH Yes. That was it. Simple. Let whoever came afterwards figure out the rest. Let them decide. He smiled and finished the painting. Charles was hungry. He got up and started for one of the restaurants near the park. Later on, when there was more time, he'd find a piece of granite and move it to the plot. He could spend his free time carving on it, copying the inscription. He would make it into a real shrine; maybe he would practice up a bit and try to carve a statue to go with the stone. Somehow, though, since things were ready and it didn't make too much difference, it seemed to Charles that he'd probably have a long time to wait. "Maybe it's just a disease, and I'm immune. I was immune to smallpox. The vaccination never took. That's probably it." He smiled. Strange, but now he wanted very much to go on living, alone or not. There were things he could do, ways to keep occupied. He wouldn't mind it so much. But he wanted more and more desperately with each passing second to retain his foothold on the tenuous path of physical existence. The tantalizing thought of "why" puzzled its way back into his mind. But it seemed less pressing now that he had almost come to the conclusion that he would live for a long time. Later, in a few days perhaps, he would think about it. In a little while he'd have plenty of opportunity for hunting down the answer. This seemed good to him, for now he thought he almost had the answer, if there were an answer. He thought he had seen the solution peering out at him from the recesses of his mind, and he didn't like the expression on its face. Better to forget. Charles reached the broad boulevard. There was a large cafe just across from him, its front window caved in by a large truck. He stumbled and almost fell as he stepped from the curb. "Look at me, nervous as a cat." He was trembling noticeably as he started across the street. "I—" He started to say something, to think something. But some hidden part of his mind clamped down, obscuring the thought, rejecting the concept. The tremor turned to a shake before he reached the far curb, and the first burst of wild pain came as he laid his shoulder against the door to the restaurant. This was the way the plague began, but—His mind quickly repressed the idea. It couldn't be the plague. He was immune! Another burst of pulsating, shattering pain crashed through his body, tearing down the defenses of his mind, putting an end of his thoughts of immunity. Colors flared before his eyes, a persistent, irresistible susurrus flooded his ears. He wanted to protest, but there was no one to listen to him. He appealed to every divinity he knew, all the time knowing it would be useless. His body, out of his voluntary control, tried to run off in all directions at once. Charles struggled to end his body's disorganized responses, to channelize all his energy into one direction. His mind came back into action. He set up his goal; everything else seemed irrelevant: he had to get back to the park, to his hermit's cave, to his long, narrow home. He couldn't die until then. Ten minutes. He was allotted ten minutes before the end. It could have been ten years or ten seconds, for now objective time meant nothing to him. It was not a matter of measuring seconds and minutes. It was a matter of forgetting time and measuring space. He concentrated on the grave; he forced his body to become an unwilling machine. While he could, he walked, forcing himself on. When his legs gave way, he crawled. When his knees buckled, he rolled. When his stomach protested, he vomited. It made no difference. Charles refused to think. Machines, especially half-broken machines, do not think; they only work. Sweating, straining, bleeding, retching, he pushed himself towards his goal, trying to add one final touch of grace and custom to the rude irrationalness of it all. His eyes gave out a few feet from the pit. He felt his way towards it. Convulsions shook his body like a cat shakes a captive mouse. He humped his body forward between the seizures, hands outstretched, searching for the grave. And then he was upon it. One arm reached out for grass, and clutched bare space instead. He was home. He gathered energy from his final reservoirs of strength for one final movement that would throw him headlong into the shallow grave. He tensed his muscles, pulled his limbs up under him and started to roll into the hole. Instantly the thought struck him with paralyzing devastation. The answer to it all poked its face out from the recesses of his mind and sapped the last bit of his energy, corroding his nerves and dying muscles. Now he knew, and the knowing was the end of it. He collapsed at the edge of the pit. Only one arm hung loosely down into it, swinging senseless in the air, pointing accusingly at the empty coffin. The world will end, not with a bang, nor with a whimper, but with the last man's anguished cry at the unreasonableness of it all. Charles screamed. The large, invisible, ovular being that hung suspended over the Empire State Building rested from its exertion. Soon it was approached by another of its kind. "It is finished?" asked the second. "Yes. Just now. I am resting." "I can feel the emptiness of it." "It was very good. Where were you?" "On the next planet out. No beauty to it at all; no system. How was yours?" "Beautiful," said the first. "It went according to the strictest semantic relationship following the purest mathematical principles. They made it easy for me." "Good." "Well, where to now?" "There's another system about four thoughts away. We're due there soon." "All right. Let's go." "What's that you have there?" "Oh, this?" replied the first. "It's a higher neural order compendium the Things here made up. It's what I used." "You can't take it with you, you know. They don't allow souvenirs." "I know." "Well?" "All right, all right. You're so good, see if you can compute the scatter probability." The first being moved imperceptably and the heavy plastoid binding of the book disappeared. The thousands of pages dropped softly, caught at the wind like hungry sails, separated, and pulled by the fingers of gravity, went their disparate ways. Here a page scuttled into a broken window of the Chrysler Building (read the names: Aabat, Aabbs, Aabbt). Here a page landed upright on the head of one of the library lions and sloughed softly to the ground (read the names: Looman, Loomana, Loomanabsky). Here another page crept in between the cracks of a pier on the riverfront, dropping gently to the caressing eddies of the water (read the names: Smith, Smitha, Smitj). And here two pages danced down into Central Park, pirouetted, promenaded, and finally came to rest against a propped-up piece of metal (read the names: Whit, Whita, Whitacomb). It was not until the dusty morning sun stirred up the breezes that they fluttered down into the shallow hole beneath, unnoticed. The writing on the metal, until then partially obscured by the papers, became legible: HERE LIES THE BODY OF THE LAST MAN ON EARTH— CHARLES J. ZZYZST GO TO HELL! | C. Humans |
What are the existing methods mentioned in the paper? | ### Introduction
Drug-drug interaction (DDI) is a situation when one drug increases or decreases the effect of another drug BIBREF0 . Adverse drug reactions may cause severe side effect, if two or more medicines were taken and their DDI were not investigated in detail. DDI is a common cause of illness, even a cause of death BIBREF1 . Thus, DDI databases for clinical medication decisions are proposed by some researchers. These databases such as SFINX BIBREF2 , KEGG BIBREF3 , CredibleMeds BIBREF4 help physicians and pharmacists avoid most adverse drug reactions. Traditional DDI databases are manually constructed according to clinical records, scientific research and drug specifications. For instance, The sentence “With combined use, clinicians should be aware, when phenytoin is added, of the potential for reexacerbation of pulmonary symptomatology due to lowered serum theophylline concentrations BIBREF5 ”, which is from a pharmacotherapy report, describe the side effect of phenytoin and theophylline's combined use. Then this information on specific medicines will be added to DDI databases. As drug-drug interactions have being increasingly found, manually constructing DDI database would consume a lot of manpower and resources. There has been many efforts to automatically extract DDIs from natural language BIBREF0 , BIBREF6 , BIBREF7 , BIBREF8 , BIBREF9 , BIBREF10 , BIBREF11 , BIBREF12 , mainly medical literature and clinical records. These works can be divided into the following categories: To avoid complex feature engineering and NLP toolkits' usage, we employ deep learning approaches for sentence comprehension as a whole. Our model takes in a sentence from biomedical literature which contains a drug pair and outputs what kind of DDI this drug pair belongs. This assists physicians refrain from improper combined use of drugs. In addition, the word and sentence level attentions are introduced to our model for better DDI predictions. We train our language comprehension model with labeled instances. Figure FIGREF5 shows partial records in DDI corpus BIBREF16 . We extract the sentence and drug pairs in the records. There are 3 drug pairs in this example thus we have 3 instances. The DDI corpus annotate each drug pair in the sentence with a DDI type. The DDI type, which is the most concerned information, is described in table TABREF4 . The details about how we train our model and extract the DDI type from text are described in the remaining sections. ### Related Work
In DDI extraction task, NLP methods or machine learning approaches are proposed by most of the work. Chowdhury BIBREF14 and Thomas et al. BIBREF11 proposed methods that use linguistic phenomenons and two-stage SVM to classify DDIs. FBK-irst BIBREF10 is a follow-on work which applies kernel method to the existing model and outperforms it. Neural network based approaches have been proposed by several works. Liu et al. BIBREF9 employ CNN for DDI extraction for the first time which outperforms the traditional machine learning based methods. Limited by the convolutional kernel size, the CNN can only extracted features of continuous 3 to 5 words rather than distant words. Liu et al. BIBREF8 proposed dependency-based CNN to handle distant but relevant words. Sahu et al. BIBREF12 proposed LSTM based DDI extraction approach and outperforms CNN based approach, since LSTM handles sentence as a sequence instead of slide windows. To conclude, Neural network based approaches have advantages of 1) less reliance on extra NLP toolkits, 2) simpler preprocessing procedure, 3) better performance than text analysis and machine learning methods. Drug-drug interaction extraction is a relation extraction task of natural language processing. Relation extraction aims to determine the relation between two given entities in a sentence. In recent years, attention mechanism and various neural networks are applied to relation extraction BIBREF17 , BIBREF18 , BIBREF19 , BIBREF20 , BIBREF21 . Convolutional deep neural network are utilized for extracting sentence level features in BIBREF19 . Then the sentence level features are concatenated with lexical level features, which are obtained by NLP toolkit WordNet BIBREF22 , followed by a multilayer perceptron (MLP) to classify the entities' relation. A fixed work is proposed by Nguyen et al. BIBREF21 . The convolutional kernel is set various size to capture more n-gram features. In addition, the word and position embedding are trained automatically instead of keeping constant as in BIBREF19 . Wang et al. BIBREF20 introduce multi-level attention mechanism to CNN in order to emphasize the keywords and ignore the non-critical words during relation detection. The attention CNN model outperforms previous state-of-the-art methods. Besides CNN, Recurrent neural network (RNN) has been applied to relation extraction as well. Zhang et al. BIBREF18 utilize long short-term memory network (LSTM), a typical RNN model, to represent sentence. The bidirectional LSTM chronologically captures the previous and future information, after which a pooling layer and MLP have been set to extract feature and classify the relation. Attention mechanism is added to bidirectional LSTM in BIBREF17 for relation extraction. An attention layer gives each memory cell a weight so that classifier can catch the principal feature for the relation detection. The Attention based bidirectional LSTM has been proven better than previous work. ### Proposed Model
In this section, we present our bidirectional recurrent neural network with multiple attention layer model. The overview of our architecture is shown in figure FIGREF15 . For a given instance, which describes the details about two or more drugs, the model represents each word as a vector in embedding layer. Then the bidirectional RNN layer generates a sentence matrix, each column vector in which is the semantic representation of the corresponding word. The word level attention layer transforms the sentence matrix to vector representation. Then sentence level attention layer generates final representation for the instance by combining several relevant sentences in view of the fact that these sentences have the same drug pair. Followed by a softmax classifier, the model classifies the drug pair in the given instance as specific DDI. ### Preprocessing
The DDI corpus contains thousands of XML files, each of which are constructed by several records. For a sentence containing INLINEFORM0 drugs, there are INLINEFORM1 drug pairs. We replace the interested two drugs with “drug1” and “drug2” while the other drugs are replaced by “durg0”, as in BIBREF9 did. This step is called drug blinding. For example, the sentence in figure FIGREF5 generates 3 instances after drug blinding: “drug1: an increased risk of hepatitis has been reported to result from combined use of drug2 and drug0”, “drug1: an increased risk of hepatitis has been reported to result from combined use of drug0 and drug2”, “drug0: an increased risk of hepatitis has been reported to result from combined use of drug1 and drug2”. The drug blinded sentences are the instances that are fed to our model. We put the sentences with the same drug pairs together as a set, since the sentence level attention layer (will be described in Section SECREF21 ) will use the sentences which contain the same drugs. ### Embedding Layer
Given an instance INLINEFORM0 which contains specified two drugs INLINEFORM1 , INLINEFORM2 , each word is embedded in a INLINEFORM3 dimensional space ( INLINEFORM4 , INLINEFORM5 are the dimension of word embedding and position embedding). The look up table function INLINEFORM6 maps a word or a relative position to a column vector. After embedding layer the sentence is represented by INLINEFORM7 , where DISPLAYFORM0 The INLINEFORM0 function is usually implemented with matrix-vector product. Let INLINEFORM1 , INLINEFORM2 denote the one-hot representation (column vector) of word and relative distance. INLINEFORM3 , INLINEFORM4 are word and position embedding query matrix. The look up functions are implemented by DISPLAYFORM0 Then the word sequence INLINEFORM0 is fed to the RNN layer. Note that the sentence will be filled with INLINEFORM1 if its length is less than INLINEFORM2 . ### Bidirectional RNN Encoding Layer
The words in the sequence are read by RNN's gated recurrent unit (GRU) one by one. The GRU takes the current word INLINEFORM0 and the previous GRU's hidden state INLINEFORM1 as input. The current GRU encodes INLINEFORM2 and INLINEFORM3 into a new hidden state INLINEFORM4 (its dimension is INLINEFORM5 , a hyperparameter), which can be regarded as informations the GRU remembered. Figure FIGREF25 shows the details in GRU. The reset gate INLINEFORM0 selectively forgets informations delivered by previous GRU. Then the hidden state becomes INLINEFORM1 . The update gate INLINEFORM2 updates the informations according to INLINEFORM3 and INLINEFORM4 . The equations below describe these procedures. Note that INLINEFORM5 stands for element wise multiplication. DISPLAYFORM0 DISPLAYFORM1 The bidirectional RNN contains forward RNN and backward RNN. Forward RNN reads sentence from INLINEFORM0 to INLINEFORM1 , generating INLINEFORM2 , INLINEFORM3 , ..., INLINEFORM4 . Backward RNN reads sentence from INLINEFORM5 to INLINEFORM6 , generating INLINEFORM7 , INLINEFORM8 , ..., INLINEFORM9 . Then the encode result of this layer is DISPLAYFORM0 We apply dropout technique in RNN layer to avoid overfitting. Each GRU have a probability (denoted by INLINEFORM0 , also a hyperparameter) of being dropped. The dropped GRU has no output and will not affect the subsequent GRUs. With bidirectional RNN and dropout technique, the input INLINEFORM1 is encoded into sentence matrix INLINEFORM2 . ### Word Level Attention
The purpose of word level attention layer is to extract sentence representation (also known as feature vector) from encoded matrix. We use word level attention instead of max pooling, since attention mechanism can determine the importance of individual encoded word in each row of INLINEFORM0 . Let INLINEFORM1 denotes the attention vector (column vector), INLINEFORM2 denotes the filter that gives each element in the row of INLINEFORM3 a weight. The following equations shows the attention operation, which is also illustrated in figure FIGREF15 . DISPLAYFORM0 DISPLAYFORM1 The softmax function takes a vector INLINEFORM0 as input and outputs a vector, DISPLAYFORM0 INLINEFORM0 denotes the feature vector captured by this layer. Several approaches BIBREF12 , BIBREF17 use this vector and softmax classifier for classification. Inspired by BIBREF23 we propose the sentence level attention to combine the information of other sentences for a improved DDI classification. ### Sentence Level Attention
The previous layers captures the features only from the given sentence. However, other sentences may contains informations that contribute to the understanding of this sentence. It is reasonable to look over other relevant instances when determine two drugs' interaction from the given sentence. In our implementation, the instances that have the same drug pair are believed to be relevant. The relevant instances set is denoted by INLINEFORM0 , where INLINEFORM1 is the sentence feature vector. INLINEFORM2 stands for how well the instance INLINEFORM3 matches its DDI INLINEFORM4 (Vector representation of a specific DDI). INLINEFORM5 is a diagonal attention matrix, multiplied by which the feature vector INLINEFORM6 can concentrate on those most representative features. DISPLAYFORM0 DISPLAYFORM1 INLINEFORM0 is the softmax result of INLINEFORM1 . The final sentence representation is decided by all of the relevant sentences' feature vector, as Equation EQREF24 shows. DISPLAYFORM0 Note that the set INLINEFORM0 is gradually growing as new sentence with the same drugs pairs is found when training. An instance INLINEFORM1 is represented by INLINEFORM2 before sentence level attention. The sentence level attention layer finds the set INLINEFORM3 , instances in which have the same drug pair as in INLINEFORM4 , and put INLINEFORM5 in INLINEFORM6 . Then the final sentence representation INLINEFORM7 is calculated in this layer. ### Classification and Training
A given sentence INLINEFORM0 is finally represented by the feature vector INLINEFORM1 . Then we feed it to a softmax classifier. Let INLINEFORM2 denotes the set of all kinds of DDI. The output INLINEFORM3 is the probabilities of each class INLINEFORM4 belongs. DISPLAYFORM0 We use cross entropy cost function and INLINEFORM0 regularization as the optimization objective. For INLINEFORM1 -th instance, INLINEFORM2 denotes the one-hot representation of it's label, where the model outputs INLINEFORM3 . The cross entropy cost is: DISPLAYFORM0 For a mini-batch INLINEFORM0 , the optimization objective is: DISPLAYFORM0 All parameters in this model is: DISPLAYFORM0 We optimize the parameters of objective function INLINEFORM0 with Adam BIBREF24 , which is a variant of mini-batch stochastic gradient descent. During each train step, the gradient of INLINEFORM1 is calculated. Then INLINEFORM2 is adjusted according to the gradient. After the end of training, we have a model that is able to predict two drugs' interactions when a sentence about these drugs is given. ### DDI Prediction
The model is trained for DDI classification. The parameters in list INLINEFORM0 are tuned during the training process. Given a new sentence with two drugs, we can use this model to classify the DDI type. The DDI prediction follows the procedure described in Section SECREF6 - SECREF26 . The given sentence is eventually represented by feature vector INLINEFORM0 . Then INLINEFORM1 is classified to a specific DDI type with a softmax classifier. In next section, we will evaluate our model's DDI prediction performance and see the advantages and shortcomings of our model. ### Datasets and Evaluation Metrics
We use the DDI corpus of the 2013 DDIExtraction challenge BIBREF16 to train and test our model. The DDIs in this corpus are classified as five types. We give the definitions of these types and their example sentences, as shown in table TABREF4 . This standard dataset is made up of training set and testing set. We use the same metrics as in other drug-drug interaction extraction literature BIBREF11 , BIBREF10 , BIBREF25 , BIBREF9 , BIBREF8 , BIBREF12 : the overall precision, recall, and F1 score on testing set. INLINEFORM0 denotes the set of {False, Mechanism, Effect, Advise, Int}. The precision and recall of each INLINEFORM1 are calculated by DISPLAYFORM0 DISPLAYFORM1 Then the overall precision, recall, and F1 score are calculated by DISPLAYFORM0 Besides, we evaluate the captured feature vectors with t-SNE BIBREF26 , a visualizing and intuitive way to map a high dimensional vector into a 2 or 3-dimensional space. If the points in a low dimensional space are easy to be split, the feature vectors are believed to be more distinguishable. ### Hyperparameter Settings and Training
We use TensorFlow BIBREF27 r0.11 to implement the proposed model. The input of each word is an ordered triple (word, relative distance from drug1, relative distance from drug2). The sentence, which is represented as a matrix, is fed to the model. The output of the model is a INLINEFORM0 -dimensional vector representing the probabilities of being corresponding DDI. It is the network, parameters, and hyperparameters which decides the output vector. The network's parameters are adjusted during training, where the hyperparameters are tuned by hand. The hyperparameters after tuning are as follows. The word embedding's dimension INLINEFORM1 , the position embedding's dimension INLINEFORM2 , the hidden state's dimension INLINEFORM3 , the probability of dropout INLINEFORM4 , other hyperparameters which are not shown here are set to TensorFlow's default values. The word embedding is initialized by pre-trained word vectors using GloVe BIBREF28 , while other parameters are initialized randomly. During each training step, a mini-batch (the mini-batch size INLINEFORM0 in our implementation) of sentences is selected from training set. The gradient of objective function is calculated for parameters updating (See Section SECREF26 ). Figure FIGREF32 shows the training process. The objective function INLINEFORM0 is declining as the training mini-batches continuously sent to the model. As the testing mini-batches, the INLINEFORM1 function is fluctuating while its overall trend is descending. The instances in testing set are not participated in training so that INLINEFORM2 function is not descending so fast. However, training and testing instances have similar distribution in sample space, causing that testing instances' INLINEFORM3 tends to be smaller along with the training process. INLINEFORM4 has inverse relationship with the performance measurement. The F1 score is getting fluctuating around a specific value after enough training steps. The reason why fluctuating range is considerable is that only a tiny part of the whole training or testing set has been calculated the F1 score. Testing the whole set during every step is time consuming and not necessary. We will evaluate the model on the whole testing set in Section SECREF47 . ### Experimental Results
We save our model every 100 step and predict all the DDIs of the instances in the testing set. These predictions' F1 score is shown in figure FIGREF40 . To demonstrate the sentence level attention layer is effective, we drop this layer and then directly use INLINEFORM0 for softmax classification (See figure FIGREF15 ). The result is shown with “RNN + dynamic word embedding + ATT” curve, which illustrates that the sentence level attention layer contributes to a more accurate model. Whether a dynamic or static word embedding is better for a DDI extraction task is under consideration. Nguyen et al. BIBREF21 shows that updating word embedding at the time of other parameters being trained makes a better performance in relation extraction task. We let the embedding be static when training, while other conditions are all the same. The “RNN + static word embedding + 2ATT” curve shows this case. We can draw a conclusion that updating the initialized word embedding trains more suitable word vectors for the task, which promotes the performance. We compare our best F1 score with other state-of-the-art approaches in table TABREF39 , which shows our model has competitive advantage in dealing with drug-drug interaction extraction. The predictions confusion matrix is shown in table TABREF46 . The DDIs other than false being classified as false makes most of the classification error. It may perform better if a classifier which can tells true and false DDI apart is trained. We leave this two-stage classifier to our future work. Another phenomenon is that the “Int” type is often classified as “Effect”. The “Int” sentence describes there exists interaction between two drugs and this information implies the two drugs' combination will have good or bed effect. That's the reason why “Int” and “Effect” are often obfuscated. To evaluate the features our model captured, we employ scikit-learn BIBREF29 's t-SNE class to map high dimensional feature vectors to 2-dimensional vectors, which can be depicted on a plane. We depict all the features of the instances in testing set, as shown in figure FIGREF41 . The RNN model using dynamic word embedding and 2 layers of attention is the most distinguishable one. Unfortunately, the classifier can not classify all the instances into correct classes. Comparing table TABREF46 with figure UID44 , both of which are from the best performed model, we can observe some conclusions. The “Int” DDIs are often misclassified as “Effect”, for the reason that some of the “Int” points are in the “Effect” cluster. The “Effect” points are too scattered so that plenty of “Effect” DDIs are classified to other types. The “Mechanism” points are gathered around two clusters, causing that most of the “mechanism” DDIs are classified to two types: “False” and “Mechanism”. In short, the visualizability of feature mapping gives better explanations for the prediction results and the quality of captured features. ### Conclusion and Future Work
To conclude, we propose a recurrent neural network with multiple attention layers to extract DDIs from biomedical text. The sentence level attention layer, which combines other sentences containing the same drugs, has been added to our model. The experiments shows that our model outperforms the state-of-the-art DDI extraction systems. Task relevant word embedding and two attention layers improved the performance to some extent. The imbalance of the classes and the ambiguity of semantics cause most of the misclassifications. We consider that instance generation using generative adversarial networks would cover the instance shortage in specific category. It is also reasonable to use distant supervision learning (which utilize other relevant material) for knowledge supplement and obtain a better performed DDI extraction system. ### Acknowledgment
This work is supported by the NSFC under Grant 61303191, 61303190, 61402504, 61103015. TABLE I THE DDI TYPES AND CORRESPONDING EXAMPLES Fig. 1. Partial records in DDI corpus Fig. 2. The bidirectional recurrent neural network with multiple attentions Fig. 3. The Gated Recurrent Unit Fig. 4. The objective function and F1 in the train process Fig. 5. The F1 scores on the whole testing set TABLE II PERFORMANCE COMPARISON WITH OTHER APPROACHES Fig. 6. The features which mapped to 2D TABLE III PREDICTION RESULTS | Chowdhury BIBREF14 and Thomas et al. BIBREF11, FBK-irst BIBREF10, Liu et al. BIBREF9, Sahu et al. BIBREF12 |
Why was Si given a symbolic gold watch by the Department of Space Exploration?
A. He had just successfully completed a dangerous space mission that they were impressed with.
B. As an apology for the difficult task he had to complete while in space.
C. He was retiring from the Department.
D. As a means to convince him to stay on with the Department and continue completing missions.
| SPACEMAN ON A SPREE BY MACK REYNOLDS Illustrated by Nodel [Transcriber's Note: This etext was produced from Worlds of Tomorrow June 1963 Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed.] What's more important—Man's conquest of space, or one spaceman's life? I They gave him a gold watch. It was meant to be symbolical, of course. In the old tradition. It was in the way of an antique, being one of the timepieces made generations past in the Alpine area of Eur-Asia. Its quaintness lay in the fact that it was wound, not electronically by power-radio, but by the actual physical movements of the bearer, a free swinging rotor keeping the mainspring at a constant tension. They also had a banquet for him, complete with speeches by such bigwigs of the Department of Space Exploration as Academician Lofting Gubelin and Doctor Hans Girard-Perregaux. There was also somebody from the government who spoke, but he was one of those who were pseudo-elected and didn't know much about the field of space travel nor the significance of Seymour Pond's retirement. Si didn't bother to remember his name. He only wondered vaguely why the cloddy had turned up at all. In common with recipients of gold watches of a score of generations before him, Si Pond would have preferred something a bit more tangible in the way of reward, such as a few shares of Variable Basic to add to his portfolio. But that, he supposed, was asking too much. The fact of the matter was, Si knew that his retiring had set them back. They hadn't figured he had enough shares of Basic to see him through decently. Well, possibly he didn't, given their standards. But Space Pilot Seymour Pond didn't have their standards. He'd had plenty of time to think it over. It was better to retire on a limited crediting, on a confoundedly limited crediting, than to take the two or three more trips in hopes of attaining a higher standard. He'd had plenty of time to figure it out, there alone in space on the Moon run, there on the Venus or Mars runs. There on the long, long haul to the Jupiter satellites, fearfully checking the symptoms of space cafard, the madness compounded of claustrophobia, monotony, boredom and free fall. Plenty of time. Time to decide that a one room mini-auto-apartment, complete with an autochair and built-in autobar, and with one wall a teevee screen, was all he needed to find contentment for a mighty long time. Possibly somebody like Doc Girard-Perregaux might be horrified at the idea of living in a mini-auto-apartment ... not realizing that to a pilot it was roomy beyond belief compared to the conning tower of a space craft. No. Even as Si listened to their speeches, accepted the watch and made a halting little talk of his own, he was grinning inwardly. There wasn't anything they could do. He had them now. He had enough Basic to keep him comfortably, by his standards, for the rest of his life. He was never going to subject himself to space cafard again. Just thinking about it, now, set the tic to going at the side of his mouth. They could count down and blast off, for all he gave a damn. The gold watch idea had been that of Lofting Gubelin, which was typical, he being in the way of a living anachronism himself. In fact, Academician Gubelin was possibly the only living man on North America who still wore spectacles. His explanation was that a phobia against having his eyes touched prohibited either surgery to remould his eyeballs and cure his myopia, or contact lenses. That was only an alibi so far as his closest associate, Hans Girard-Perregaux, was concerned. Doctor Girard-Perregaux was convinced Gubelin would have even worn facial hair, had he but a touch more courage. Gubelin longed for yesteryear, a seldom found phenomenon under the Ultrawelfare State. Slumped in an autochair in the escape room of his Floridian home, Lofting Gubelin scowled at his friend. He said, acidly, "Any more bright schemes, Hans? I presume you now acknowledge that appealing to the cloddy's patriotism, sentiment and desire for public acclaim have miserably failed." Girard-Perregaux said easily, "I wouldn't call Seymour Pond a cloddy. In his position, I am afraid I would do the same thing he has." "That's nonsense, Hans. Zoroaster! Either you or I would gladly take Pond's place were we capable of performing the duties for which he has been trained. There aren't two men on North America—there aren't two men in the world!—who better realize the urgency of continuing our delving into space." Gubelin snapped his fingers. "Like that, either of us would give our lives to prevent man from completely abandoning the road to his destiny." His friend said drily, "Either of us could have volunteered for pilot training forty years ago, Lofting. We didn't." "At that time there wasn't such a blistering percentage of funkers throughout this whole blistering Ultrawelfare State! Who could foresee that eventually our whole program would face ending due to lack of courageous young men willing to take chances, willing to face adventure, willing to react to the stimulus of danger in the manner our ancestors did?" Girard-Perregaux grunted his sarcasm and dialed a glass of iced tea and tequila. He said, "Nevertheless, both you and I conform with the present generation in finding it far more pleasant to follow one's way of life in the comfort of one's home than to be confronted with the unpleasantness of facing nature's dangers in more adventurous pastimes." Gubelin, half angry at his friend's argument, leaned forward to snap rebuttal, but the other was wagging a finger at him negatively. "Face reality, Lofting. Don't require or expect from Seymour Pond more than is to be found there. He is an average young man. Born in our Ultrawelfare State, he was guaranteed his fundamental womb-to-tomb security by being issued that minimum number of Basic shares in our society that allows him an income sufficient to secure the food, clothing, shelter, medical care and education to sustain a low level of subsistence. Percentages were against his ever being drafted into industry. Automation being what it is, only a fraction of the population is ever called up. But Pond was. His industrial aptitude dossier revealed him a possible candidate for space pilot, and it was you yourself who talked him into taking the training ... pointing out the more pragmatic advantages such as complete retirement after but six trips, added shares of Basic so that he could enjoy a more comfortable life than most and the fame that would accrue to him as one of the very few who still participate in travel to the planets. Very well. He was sold. Took his training, which, of course, required long years of drudgery to him. Then, performing his duties quite competently, he made his six trips. He is now legally eligible for retirement. He was drafted into the working force reserves, served his time, and is now free from toil for the balance of his life. Why should he listen to our pleas for a few more trips?" "But has he no spirit of adventure? Has he no feeling for...." Girard-Perregaux was wagging his finger again, a gesture that, seemingly mild though it was, had an astonishing ability to break off the conversation of one who debated with the easy-seeming, quiet spoken man. He said, "No, he hasn't. Few there are who have, nowadays. Man has always paid lip service to adventure, hardships and excitement, but in actuality his instincts, like those of any other animal, lead him to the least dangerous path. Today we've reached the point where no one need face danger—ever. There are few who don't take advantage of the fact. Including you and me, Lofting, and including Seymour Pond." His friend and colleague changed subjects abruptly, impatiently. "Let's leave this blistering jabber about Pond's motivation and get to the point. The man is the only trained space pilot in the world. It will take months, possibly more than a year, to bring another novitiate pilot to the point where he can safely be trusted to take our next explorer craft out. Appropriations for our expeditions have been increasingly hard to come by—even though in our minds, Hans, we are near important breakthroughs, breakthroughs which might possibly so spark the race that a new dream to push man out to the stars will take hold of us. If it is admitted that our organization has degenerated to the point that we haven't a single pilot, then it might well be that the Economic Planning Board, and especially those cloddies on Appropriations, will terminate the whole Department of Space Exploration." "So...." Girard-Perregaux said gently. "So some way we've got to bring Seymour Pond out of his retirement!" "Now we are getting to matters." Girard-Perregaux nodded his agreement. Looking over the rim of his glass, his eyes narrowed in thought as his face took on an expression of Machiavellianism. "And do not the ends justify the means?" Gubelin blinked at him. The other chuckled. "The trouble with you, Lofting, is that you have failed to bring history to bear on our problem. Haven't you ever read of the sailor and his way of life?" "Sailor? What in the name of the living Zoroaster has the sailor got to do with it?" "You must realize, my dear Lofting, that our Si Pond is nothing more than a latter-day sailor, with many of the problems and view-points, tendencies and weaknesses of the voyager of the past. Have you never heard of the seaman who dreamed of returning to the village of his birth and buying a chicken farm or some such? All the long months at sea—and sometimes the tramp freighters or whaling craft would be out for years at a stretch before returning to home port—he would talk of his retirement and his dream. And then? Then in port, it would be one short drink with the boys, before taking his accumulated pay and heading home. The one short drink would lead to another. And morning would find him, drunk, rolled, tattooed and possibly sleeping it off in jail. So back to sea he'd have to go." Gubelin grunted bitterly. "Unfortunately, our present-day sailor can't be separated from his money quite so easily. If he could, I'd personally be willing to lure him down some dark alley, knock him over the head and roll him myself. Just to bring him back to his job again." He brought his wallet from his pocket, and flicked it open to his universal credit card. "The ultimate means of exchange," he grunted. "Nobody can spend your money, but you, yourself. Nobody can steal it, nobody can, ah, con you out of it. Just how do you expect to sever our present-day sailor and his accumulated nest egg?" The other chuckled again. "It is simply a matter of finding more modern methods, my dear chap." II Si Pond was a great believer in the institution of the spree. Any excuse would do. Back when he had finished basic education at the age of twenty-five and was registered for the labor draft, there hadn't been a chance in a hundred that he'd have the bad luck to have his name pulled. But when it had been, Si had celebrated. When he had been informed that his physical and mental qualifications were such that he was eligible for the most dangerous occupation in the Ultrawelfare State and had been pressured into taking training for space pilot, he had celebrated once again. Twenty-two others had taken the training with him, and only he and Rod Cameroon had passed the finals. On this occasion, he and Rod had celebrated together. It had been quite a party. Two weeks later, Rod had burned on a faulty take-off on what should have been a routine Moon run. Each time Si returned from one of his own runs, he celebrated. A spree, a bust, a bat, a wing-ding, a night on the town. A commemoration of dangers met and passed. Now it was all over. At the age of thirty he was retired. Law prevented him from ever being called up for contributing to the country's labor needs again. And he most certainly wasn't going to volunteer. He had taken his schooling much as had his contemporaries. There wasn't any particular reason for trying to excell. You didn't want to get the reputation for being a wise guy, or a cloddy either. Just one of the fellas. You could do the same in life whether you really studied or not. You had your Inalienable Basic stock, didn't you? What else did you need? It had come as a surprise when he'd been drafted for the labor force. In the early days of the Ultrawelfare State, they had made a mistake in adapting to the automation of the second industrial revolution. They had attempted to give everyone work by reducing the number of working hours in the day, and the number of working days in the week. It finally became ludicrous when employees of industry were working but two days a week, two hours a day. In fact, it got chaotic. It became obvious that it was more practical to have one worker putting in thirty-five hours a week and getting to know his job well, than it was to have a score of employees, each working a few hours a week and none of them ever really becoming efficient. The only fair thing was to let the technologically unemployed remain unemployed, with their Inalienable Basic stock as the equivalent of unemployment insurance, while the few workers still needed put in a reasonable number of hours a day, a reasonable number of weeks a year and a reasonable number of years in a life time. When new employees were needed, a draft lottery was held. All persons registered in the labor force participated. If you were drawn, you must need serve. The dissatisfaction those chosen might feel at their poor luck was offset by the fact that they were granted additional Variable Basic shares, according to the tasks they fulfilled. Such shares could be added to their portfolios, the dividends becoming part of their current credit balance, or could be sold for a lump sum on the market. Yes, but now it was all over. He had his own little place, his own vacuum-tube vehicle and twice the amount of shares of Basic that most of his fellow citizens could boast. Si Pond had it made. A spree was obviously called for. He was going to do this one right. This was the big one. He'd accumulated a lot of dollars these past few months and he intended to blow them, or at least a sizeable number of them. His credit card was burning a hole in his pocket, as the expression went. However, he wasn't going to rush into things. This had to be done correctly. Too many a spree was played by ear. You started off with a few drinks, fell in with some second rate mopsy and usually wound up in a third rate groggery where you spent just as much as though you'd been in the classiest joint in town. Came morning and you had nothing to show for all the dollars that had been spent but a rum-head. Thus, Si was vaguely aware, it had always been down through the centuries since the Phoenecian sailor, back from his year-long trip to the tin mines of Cornwall, blew his hard earned share of the voyage's profits in a matter of days in the wine shops of Tyre. Nobody gets quite so little for his money as that loneliest of all workers, he who must leave his home for distant lands, returning only periodically and usually with the salary of lengthy, weary periods of time to be spent hurriedly in an attempt to achieve the pleasure and happiness so long denied him. Si was going to do it differently this time. Nothing but the best. Wine, women, song, food, entertainment. The works. But nothing but the best. To start off, he dressed with great care in the honorable retirement-rank suit he had so recently purchased. His space pin he attached carefully to the lapel. That was a good beginning, he decided. A bit of prestige didn't hurt you when you went out on the town. In the Ultrawelfare State hardly one person in a hundred actually ever performed anything of value to society. The efforts of most weren't needed. Those few who did contribute were awarded honors, decorations, titles. Attired satisfactorily, Si double-checked to see that his credit card was in his pocket. As an after-thought, he went over to the auto-apartment's teevee-phone, flicked it on, held the card to the screen and said, "Balance check, please." In a moment, the teevee-phone's robot voice reported, "Ten shares of Inalienable Basic. Twelve shares of Variable Basic, current value, four thousand, two hundred and thirty-three dollars and sixty-two cents apiece. Current cash credit, one thousand and eighty-four dollars." The screen went dead. One thousand and eighty-four dollars. That was plenty. He could safely spend as much as half of it, if the spree got as lively as he hoped it would. His monthly dividends were due in another week or so, and he wouldn't have to worry about current expenses. Yes, indeedy, Si Pond was as solvent as he had ever been in his thirty years. He opened the small, closet-like door which housed his vacuum-tube two-seater, and wedged himself into the small vehicle. He brought down the canopy, dropped the pressurizer and considered the dial. Only one place really made sense. The big city. He considered for a moment, decided against the boroughs of Baltimore and Boston, and selected Manhattan instead. He had the resources. He might as well do it up brown. He dialed Manhattan and felt the sinking sensation that presaged his car's dropping to tube level. While it was being taken up by the robot controls, being shuttled here and there preparatory to the shot to his destination, he dialed the vehicle's teevee-phone for information on the hotels of the island of the Hudson. He selected a swank hostelry he'd read about and seen on the teevee casts of society and celebrity gossip reporters, and dialed it on the car's destination dial. "Nothing too good for ex-Space Pilot Si Pond," he said aloud. The car hesitated for a moment, that brief hesitation before the shot, and Si took the involuntary breath from which only heroes could refrain. He sank back slowly into the seat. Moments passed, and the direction of the pressure was reversed. Manhattan. The shuttling began again, and one or two more traversing sub-shots. Finally, the dash threw a green light and Si opened the canopy and stepped into his hotel room. A voice said gently, "If the quarters are satisfactory, please present your credit card within ten minutes." Si took his time. Not that he really needed it. It was by far the most swank suite he had ever seen. One wall was a window of whatever size the guest might desire and Si touched the control that dilated it to the full. His view opened in such wise that he could see both the Empire State Building Museum and the Hudson. Beyond the river stretched the all but endless city which was Greater Metropolis. He didn't take the time to flick on the menu, next to the auto-dining table, nor to check the endless potables on the autobar list. All that, he well knew, would be superlative. Besides, he didn't plan to dine or do much drinking in his suite. He made a mock leer. Not unless he managed to acquire some feminine companionship, that was. He looked briefly into the swimming pool and bath, then flopped himself happily onto the bed. It wasn't up to the degree of softness he presently desired, and he dialed the thing to the ultimate in that direction so that with a laugh he sank almost out of sight into the mattress. He came back to his feet, gave his suit a quick patting so that it fell into press and, taking his credit card from his pocket, put it against the teevee-phone screen and pressed the hotel button so that registration could be completed. For a moment he stood in the center of the floor, in thought. Take it easy, Si Pond, take it all easy, this time. No throwing his dollars around in second-class groggeries, no eating in automated luncheterias. This time, be it the only time in his life, he was going to frolic in the grand manner. No cloddy was Si Pond. He decided a drink was in order to help him plan his strategy. A drink at the hotel's famous Kudos Room where celebrities were reputed to be a dime a dozen. He left the suite and stepped into one of the elevators. He said, "Kudos Room." The auto-elevator murmured politely, "Yes, sir, the Kudos Room." At the door to the famous rendezvous of the swankiest set, Si paused a moment and looked about. He'd never been in a place like this, either. However, he stifled his first instinct to wonder about what this was going to do to his current credit balance with an inner grin and made his way to the bar. There was actually a bartender. Si Pond suppressed his astonishment and said, offhand, attempting an air of easy sophistication, "Slivovitz Sour." "Yes, sir." The drinks in the Kudos Room might be concocted by hand, but Si noticed they had the routine teevee screens built into the bar for payment. He put his credit card on the screen immediately before him when the drink came, and had to quell his desire to dial for a balance check, so as to be able to figure out what the Sour had cost him. Well, this was something like it. This was the sort of thing he'd dreamed about, out there in the great alone, seated in the confining conning tower of his space craft. He sipped at the drink, finding it up to his highest expectations, and then swiveled slightly on his stool to take a look at the others present. To his disappointment, there were no recognizable celebrities. None that he placed, at least—top teevee stars, top politicians of the Ultrawelfare State or Sports personalities. He turned back to his drink and noticed, for the first time, the girl who occupied the stool two down from him. Si Pond blinked. He blinked and then swallowed. " Zo-ro-as-ter ," he breathed. She was done in the latest style from Shanghai, even to the point of having cosmetically duplicated the Mongolian fold at the corners of her eyes. Every pore, but every pore, was in place. She sat with the easy grace of the Orient, so seldom found in the West. His stare couldn't be ignored. She looked at him coldly, turned to the bartender and murmured, "A Far Out Cooler, please, Fredric." Then deliberately added, "I thought the Kudos Room was supposed to be exclusive." There was nothing the bartender could say to that, and he went about building the drink. Si cleared his throat. "Hey," he said, "how about letting this one be on me?" Her eyebrows, which had been plucked and penciled to carry out her Oriental motif, rose. "Really!" she said, drawing it out. The bartender said hurriedly, "I beg your pardon, sir...." The girl, her voice suddenly subtly changed, said, "Why, isn't that a space pin?" Si, disconcerted by the sudden reversal, said, "Yeah ... sure." "Good Heavens, you're a spaceman?" "Sure." He pointed at the lapel pin. "You can't wear one unless you been on at least a Moon run." She was obviously both taken back and impressed. "Why," she said, "you're Seymour Pond, the pilot. I tuned in on the banquet they gave you." Si, carrying his glass, moved over to the stool next to her. "Call me Si," he said. "Everybody calls me Si." She said, "I'm Natalie. Natalie Paskov. Just Natalie. Imagine meeting Seymour Pond. Just sitting down next to him at a bar. Just like that." "Si," Si said, gratified. Holy Zoroaster, he'd never seen anything like this rarified pulchritude. Maybe on teevee, of course, one of the current sex symbols, but never in person. "Call me Si," he said again. "I been called Si so long, I don't even know who somebody's talking to if they say Seymour." "I cried when they gave you that antique watch," she said, her tone such that it was obvious she hadn't quite adjusted as yet to having met him. Si Pond was surprised. "Cried?" he said. "Well, why? I was kind of bored with the whole thing. But old Doc Gubelin, I used to work under him in the Space Exploration department, he was hot for it." " Academician Gubelin?" she said. "You just call him Doc ?" Si was expansive. "Why, sure. In the Space Department we don't have much time for formality. Everybody's just Si, and Doc, and Jim. Like that. But how come you cried?" She looked down into the drink the bartender had placed before her, as though avoiding his face. "I ... I suppose it was that speech Doctor Girard-Perregaux made. There you stood, so fine and straight in your space-pilot uniform, the veteran of six exploration runs to the planets...." "Well," Si said modestly, "two of my runs were only to the Moon." "... and he said all those things about man's conquest of space. And the dream of the stars which man has held so long. And then the fact that you were the last of the space pilots. The last man in the whole world trained to pilot a space craft. And here you were, retiring." Si grunted. "Yeah. That's all part of the Doc's scheme to get me to take on another three runs. They're afraid the whole department'll be dropped by the Appropriations Committee on this here Economic Planning Board. Even if they can find some other patsy to train for the job, it'd take maybe a year before you could even send him on a Moon hop. So old man Gubelin, and Girard-Perregaux too, they're both trying to pressure me into more trips. Otherwise they got a Space Exploration Department, with all the expense and all, but nobody to pilot their ships. It's kind of funny, in a way. You know what one of those spaceships costs?" "Funny?" she said. "Why, I don't think it's funny at all." Si said, "Look, how about another drink?" Natalie Paskov said, "Oh, I'd love to have a drink with you, Mr...." "Si," Si said. He motioned to the bartender with a circular twist of the hand indicating their need for two more of the same. "How come you know so much about it? You don't meet many people who are interested in space any more. In fact, most people are almost contemptuous, like. Think it's kind of a big boondoggle deal to help use up a lot of materials and all and keep the economy going." Natalie said earnestly, "Why, I've been a space fan all my life. I've read all about it. Have always known the names of all the space pilots and everything about them, ever since I was a child. I suppose you'd say I have the dream that Doctor Girard-Perregaux spoke about." Si chuckled. "A real buff, eh? You know, it's kind of funny. I was never much interested in it. And I got a darn sight less interested after my first run and I found out what space cafard was." She frowned. "I don't believe I know much about that." Sitting in the Kudos Room with the most beautiful girl to whom he had ever talked, Si could be nonchalant about the subject. "Old Gubelin keeps that angle mostly hushed up and out of the magazine and newspaper articles. Says there's enough adverse publicity about space exploration already. But at this stage of the game when the whole ship's crammed tight with this automatic scientific apparatus and all, there's precious little room in the conning tower and you're the only man aboard. The Doc says later on when ships are bigger and there's a whole flock of people aboard, there won't be any such thing as space cafard, but...." Of a sudden the right side of Si Pond's mouth began to tic and he hurriedly took up his drink and knocked it back. | C. He was retiring from the Department. |
What is the name of the space cruiser that the Groacians are hiding?
A. The Terran
B. The Territory
C. The Terror
D. The Terrific
| THE MADMAN FROM EARTH BY KEITH LAUMER You don't have to be crazy to be an earth diplomat—but on Groac it sure helps! [Transcriber's Note: This etext was produced from Worlds of If Science Fiction, March 1962. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed.] I "The Consul for the Terrestrial States," Retief said, "presents his compliments, et cetera, to the Ministry of Culture of the Groacian Autonomy, and with reference to the Ministry's invitation to attend a recital of interpretive grimacing, has the honor to express regret that he will be unable—" "You can't turn this invitation down," Administrative Assistant Meuhl said flatly. "I'll make that 'accepts with pleasure'." Retief exhaled a plume of cigar smoke. "Miss Meuhl," he said, "in the past couple of weeks I've sat through six light-concerts, four attempts at chamber music, and god knows how many assorted folk-art festivals. I've been tied up every off-duty hour since I got here—" "You can't offend the Groaci," Miss Meuhl said sharply. "Consul Whaffle would never have been so rude." "Whaffle left here three months ago," Retief said, "leaving me in charge." "Well," Miss Meuhl said, snapping off the dictyper. "I'm sure I don't know what excuse I can give the Minister." "Never mind the excuses," Retief said. "Just tell him I won't be there." He stood up. "Are you leaving the office?" Miss Meuhl adjusted her glasses. "I have some important letters here for your signature." "I don't recall dictating any letters today, Miss Meuhl," Retief said, pulling on a light cape. "I wrote them for you. They're just as Consul Whaffle would have wanted them." "Did you write all Whaffle's letters for him, Miss Meuhl?" "Consul Whaffle was an extremely busy man," Miss Meuhl said stiffly. "He had complete confidence in me." "Since I'm cutting out the culture from now on," Retief said, "I won't be so busy." "Well!" Miss Meuhl said. "May I ask where you'll be if something comes up?" "I'm going over to the Foreign Office Archives." Miss Meuhl blinked behind thick lenses. "Whatever for?" Retief looked thoughtfully at Miss Meuhl. "You've been here on Groac for four years, Miss Meuhl. What was behind the coup d'etat that put the present government in power?" "I'm sure I haven't pried into—" "What about that Terrestrial cruiser? The one that disappeared out this way about ten years back?" "Mr. Retief, those are just the sort of questions we avoid with the Groaci. I certainly hope you're not thinking of openly intruding—" "Why?" "The Groaci are a very sensitive race. They don't welcome outworlders raking up things. They've been gracious enough to let us live down the fact that Terrestrials subjected them to deep humiliation on one occasion." "You mean when they came looking for the cruiser?" "I, for one, am ashamed of the high-handed tactics that were employed, grilling these innocent people as though they were criminals. We try never to reopen that wound, Mr. Retief." "They never found the cruiser, did they?" "Certainly not on Groac." Retief nodded. "Thanks, Miss Meuhl," he said. "I'll be back before you close the office." Miss Meuhl's face was set in lines of grim disapproval as he closed the door. The pale-featured Groacian vibrated his throat-bladder in a distressed bleat. "Not to enter the Archives," he said in his faint voice. "The denial of permission. The deep regret of the Archivist." "The importance of my task here," Retief said, enunciating the glottal dialect with difficulty. "My interest in local history." "The impossibility of access to outworlders. To depart quietly." "The necessity that I enter." "The specific instructions of the Archivist." The Groacian's voice rose to a whisper. "To insist no longer. To give up this idea!" "OK, Skinny, I know when I'm licked," Retief said in Terran. "To keep your nose clean." Outside, Retief stood for a moment looking across at the deeply carved windowless stucco facades lining the street, then started off in the direction of the Terrestrial Consulate General. The few Groacians on the street eyed him furtively, veered to avoid him as he passed. Flimsy high-wheeled ground cars puffed silently along the resilient pavement. The air was clean and cool. At the office, Miss Meuhl would be waiting with another list of complaints. Retief studied the carving over the open doorways along the street. An elaborate one picked out in pinkish paint seemed to indicate the Groacian equivalent of a bar. Retief went in. A Groacian bartender was dispensing clay pots of alcoholic drink from the bar-pit at the center of the room. He looked at Retief and froze in mid-motion, a metal tube poised over a waiting pot. "To enjoy a cooling drink," Retief said in Groacian, squatting down at the edge of the pit. "To sample a true Groacian beverage." "To not enjoy my poor offerings," the Groacian mumbled. "A pain in the digestive sacs; to express regret." "To not worry," Retief said, irritated. "To pour it out and let me decide whether I like it." "To be grappled in by peace-keepers for poisoning of—foreigners." The barkeep looked around for support, found none. The Groaci customers, eyes elsewhere, were drifting away. "To get the lead out," Retief said, placing a thick gold-piece in the dish provided. "To shake a tentacle." "The procuring of a cage," a thin voice called from the sidelines. "The displaying of a freak." Retief turned. A tall Groacian vibrated his mandibles in a gesture of contempt. From his bluish throat coloration, it was apparent the creature was drunk. "To choke in your upper sac," the bartender hissed, extending his eyes toward the drunk. "To keep silent, litter-mate of drones." "To swallow your own poison, dispenser of vileness," the drunk whispered. "To find a proper cage for this zoo-piece." He wavered toward Retief. "To show this one in the streets, like all freaks." "Seen a lot of freaks like me, have you?" Retief asked, interestedly. "To speak intelligibly, malodorous outworlder," the drunk said. The barkeep whispered something, and two customers came up to the drunk, took his arms and helped him to the door. "To get a cage!" the drunk shrilled. "To keep the animals in their own stinking place." "I've changed my mind," Retief said to the bartender. "To be grateful as hell, but to have to hurry off now." He followed the drunk out the door. The other Groaci released him, hurried back inside. Retief looked at the weaving alien. "To begone, freak," the Groacian whispered. "To be pals," Retief said. "To be kind to dumb animals." "To have you hauled away to a stockyard, ill-odored foreign livestock." "To not be angry, fragrant native," Retief said. "To permit me to chum with you." "To flee before I take a cane to you!" "To have a drink together—" "To not endure such insolence!" The Groacian advanced toward Retief. Retief backed away. "To hold hands," Retief said. "To be palsy-walsy—" The Groacian reached for him, missed. A passer-by stepped around him, head down, scuttled away. Retief backed into the opening to a narrow crossway and offered further verbal familiarities to the drunken local, who followed, furious. Retief backed, rounded a corner into a narrow alley-like passage, deserted, silent ... except for the following Groacian. Retief stepped around him, seized his collar and yanked. The Groacian fell on his back. Retief stood over him. The downed native half-rose; Retief put a foot against his chest and pushed. "To not be going anywhere for a few minutes," Retief said. "To stay right here and have a nice long talk." II "There you are!" Miss Meuhl said, eyeing Retief over her lenses. "There are two gentlemen waiting to see you. Groacian gentlemen." "Government men, I imagine. Word travels fast." Retief pulled off his cape. "This saves me the trouble of paying another call at the Foreign Ministry." "What have you been doing? They seem very upset, I don't mind telling you." "I'm sure you don't. Come along. And bring an official recorder." Two Groaci wearing heavy eye-shields and elaborate crest ornaments indicative of rank rose as Retief entered the room. Neither offered a courteous snap of the mandibles, Retief noted. They were mad, all right. "I am Fith, of the Terrestrial Desk, Ministry of Foreign Affairs, Mr. Consul," the taller Groacian said, in lisping Terran. "May I present Shluh, of the Internal Police?" "Sit down, gentlemen," Retief said. They resumed their seats. Miss Meuhl hovered nervously, then sat on the edge of a comfortless chair. "Oh, it's such a pleasure—" she began. "Never mind that," Retief said. "These gentlemen didn't come here to sip tea today." "So true," Fith said. "Frankly, I have had a most disturbing report, Mr. Consul. I shall ask Shluh to recount it." He nodded to the police chief. "One hour ago," The Groacian said, "a Groacian national was brought to hospital suffering from serious contusions. Questioning of this individual revealed that he had been set upon and beaten by a foreigner. A Terrestrial, to be precise. Investigation by my department indicates that the description of the culprit closely matches that of the Terrestrial Consul." Miss Meuhl gasped audibly. "Have you ever heard," Retief said, looking steadily at Fith, "of a Terrestrial cruiser, the ISV Terrific , which dropped from sight in this sector nine years ago?" "Really!" Miss Meuhl exclaimed, rising. "I wash my hands—" "Just keep that recorder going," Retief snapped. "I'll not be a party—" "You'll do as you're told, Miss Meuhl," Retief said quietly. "I'm telling you to make an official sealed record of this conversation." Miss Meuhl sat down. Fith puffed out his throat indignantly. "You reopen an old wound, Mr. Consul. It reminds us of certain illegal treatment at Terrestrial hands—" "Hogwash," Retief said. "That tune went over with my predecessors, but it hits a sour note with me." "All our efforts," Miss Meuhl said, "to live down that terrible episode! And you—" "Terrible? I understand that a Terrestrial task force stood off Groac and sent a delegation down to ask questions. They got some funny answers, and stayed on to dig around a little. After a week they left. Somewhat annoying to the Groaci, maybe—at the most. If they were innocent." "IF!" Miss Meuhl burst out. "If, indeed!" Fith said, his weak voice trembling. "I must protest your—" "Save the protests, Fith. You have some explaining to do. And I don't think your story will be good enough." "It is for you to explain! This person who was beaten—" "Not beaten. Just rapped a few times to loosen his memory." "Then you admit—" "It worked, too. He remembered lots of things, once he put his mind to it." Fith rose; Shluh followed suit. "I shall ask for your immediate recall, Mr. Consul. Were it not for your diplomatic immunity, I should do more—" "Why did the government fall, Fith? It was just after the task force paid its visit, and before the arrival of the first Terrestrial diplomatic mission." "This is an internal matter!" Fith cried, in his faint Groacian voice. "The new regime has shown itself most amiable to you Terrestrials. It has outdone itself—" "—to keep the Terrestrial consul and his staff in the dark," Retief said. "And the same goes for the few terrestrial businessmen you've visaed. This continual round of culture; no social contacts outside the diplomatic circle; no travel permits to visit out-lying districts, or your satellite—" "Enough!" Fith's mandibles quivered in distress. "I can talk no more of this matter—" "You'll talk to me, or there'll be a task force here in five days to do the talking," Retief said. "You can't!" Miss Meuhl gasped. Retief turned a steady look on Miss Meuhl. She closed her mouth. The Groaci sat down. "Answer me this one," Retief said, looking at Shluh. "A few years back—about nine, I think—there was a little parade held here. Some curious looking creatures were captured. After being securely caged, they were exhibited to the gentle Groaci public. Hauled through the streets. Very educational, no doubt. A highly cultural show. "Funny thing about these animals. They wore clothes. They seemed to communicate with each other. Altogether it was a very amusing exhibit. "Tell me, Shluh, what happened to those six Terrestrials after the parade was over?" Fith made a choked noise and spoke rapidly to Shluh in Groacian. Shluh retracted his eyes, shrank down in his chair. Miss Meuhl opened her mouth, closed it and blinked rapidly. "How did they die?" Retief snapped. "Did you murder them, cut their throats, shoot them or bury them alive? What amusing end did you figure out for them? Research, maybe? Cut them open to see what made them yell...." "No!" Fith gasped. "I must correct this terrible false impression at once." "False impression, hell," Retief said. "They were Terrans! A simple narco-interrogation would get that out of any Groacian who saw the parade." "Yes," Fith said weakly. "It is true, they were Terrestrials. But there was no killing." "They're alive?" "Alas, no. They ... died." Miss Meuhl yelped faintly. "I see," Retief said. "They died." "We tried to keep them alive, of course. But we did not know what foods—" "Didn't take the trouble to find out, either, did you?" "They fell ill," Fith said. "One by one...." "We'll deal with that question later," Retief said. "Right now, I want more information. Where did you get them? Where did you hide the ship? What happened to the rest of the crew? Did they 'fall ill' before the big parade?" "There were no more! Absolutely, I assure you!" "Killed in the crash landing?" "No crash landing. The ship descended intact, east of the city. The ... Terrestrials ... were unharmed. Naturally, we feared them. They were strange to us. We had never before seen such beings." "Stepped off the ship with guns blazing, did they?" "Guns? No, no guns—" "They raised their hands, didn't they? Asked for help. You helped them; helped them to death." "How could we know?" Fith moaned. "How could you know a flotilla would show up in a few months looking for them, you mean? That was a shock, wasn't it? I'll bet you had a brisk time of it hiding the ship, and shutting everybody up. A close call, eh?" "We were afraid," Shluh said. "We are a simple people. We feared the strange creatures from the alien craft. We did not kill them, but we felt it was as well they ... did not survive. Then, when the warships came, we realized our error. But we feared to speak. We purged our guilty leaders, concealed what had happened, and ... offered our friendship. We invited the opening of diplomatic relations. We made a blunder, it is true, a great blunder. But we have tried to make amends...." "Where is the ship?" "The ship?" "What did you do with it? It was too big to just walk off and forget. Where is it?" The two Groacians exchanged looks. "We wish to show our contrition," Fith said. "We will show you the ship." "Miss Meuhl," Retief said. "If I don't come back in a reasonable length of time, transmit that recording to Regional Headquarters, sealed." He stood, looked at the Groaci. "Let's go," he said. Retief stooped under the heavy timbers shoring the entry to the cavern. He peered into the gloom at the curving flank of the space-burned hull. "Any lights in here?" he asked. A Groacian threw a switch. A weak bluish glow sprang up. Retief walked along the raised wooden catwalk, studying the ship. Empty emplacements gaped below lensless scanner eyes. Littered decking was visible within the half-open entry port. Near the bow the words 'IVS Terrific B7 New Terra' were lettered in bright chrome duralloy. "How did you get it in here?" Retief asked. "It was hauled here from the landing point, some nine miles distant," Fith said, his voice thinner than ever. "This is a natural crevasse. The vessel was lowered into it and roofed over." "How did you shield it so the detectors didn't pick it up?" "All here is high-grade iron ore," Fith said, waving a member. "Great veins of almost pure metal." Retief grunted. "Let's go inside." Shluh came forward with a hand-lamp. The party entered the ship. Retief clambered up a narrow companionway, glanced around the interior of the control compartment. Dust was thick on the deck, the stanchions where acceleration couches had been mounted, the empty instrument panels, the litter of sheared bolts, scraps of wire and paper. A thin frosting of rust dulled the exposed metal where cutting torches had sliced away heavy shielding. There was a faint odor of stale bedding. "The cargo compartment—" Shluh began. "I've seen enough," Retief said. Silently, the Groacians led the way back out through the tunnel and into the late afternoon sunshine. As they climbed the slope to the steam car, Fith came to Retief's side. "Indeed, I hope that this will be the end of this unfortunate affair," he said. "Now that all has been fully and honestly shown—" "You can skip all that," Retief said. "You're nine years late. The crew was still alive when the task force called, I imagine. You killed them—or let them die—rather than take the chance of admitting what you'd done." "We were at fault," Fith said abjectly. "Now we wish only friendship." "The Terrific was a heavy cruiser, about twenty thousand tons." Retief looked grimly at the slender Foreign Office official. "Where is she, Fith? I won't settle for a hundred-ton lifeboat." Fith erected his eye stalks so violently that one eye-shield fell off. "I know nothing of ... of...." He stopped. His throat vibrated rapidly as he struggled for calm. "My government can entertain no further accusations, Mr. Consul," he said at last. "I have been completely candid with you, I have overlooked your probing into matters not properly within your sphere of responsibility. My patience is at an end." "Where is that ship?" Retief rapped out. "You never learn, do you? You're still convinced you can hide the whole thing and forget it. I'm telling you you can't." "We return to the city now," Fith said. "I can do no more." "You can and you will, Fith," Retief said. "I intend to get to the truth of this matter." Fith spoke to Shluh in rapid Groacian. The police chief gestured to his four armed constables. They moved to ring Retief in. Retief eyed Fith. "Don't try it," he said. "You'll just get yourself in deeper." Fith clacked his mandibles angrily, eye stalks canted aggressively toward the Terrestrial. "Out of deference to your diplomatic status, Terrestrial, I shall ignore your insulting remarks," Fith said in his reedy voice. "Let us now return to the city." Retief looked at the four policemen. "I see your point," he said. Fith followed him into the car, sat rigidly at the far end of the seat. "I advise you to remain very close to your consulate," Fith said. "I advise you to dismiss these fancies from your mind, and to enjoy the cultural aspects of life at Groac. Especially, I should not venture out of the city, or appear overly curious about matters of concern only to the Groacian government." In the front seat, Shluh looked straight ahead. The loosely-sprung vehicle bobbed and swayed along the narrow highway. Retief listened to the rhythmic puffing of the motor and said nothing. III "Miss Meuhl," Retief said, "I want you to listen carefully to what I'm going to tell you. I have to move rapidly now, to catch the Groaci off guard." "I'm sure I don't know what you're talking about," Miss Meuhl snapped, her eyes sharp behind the heavy lenses. "If you'll listen, you may find out," Retief said. "I have no time to waste, Miss Meuhl. They won't be expecting an immediate move—I hope—and that may give me the latitude I need." "You're still determined to make an issue of that incident!" Miss Meuhl snorted. "I really can hardly blame the Groaci. They are not a sophisticated race; they had never before met aliens." "You're ready to forgive a great deal, Miss Meuhl. But it's not what happened nine years ago I'm concerned with. It's what's happening now. I've told you that it was only a lifeboat the Groaci have hidden out. Don't you understand the implication? That vessel couldn't have come far. The cruiser itself must be somewhere near by. I want to know where!" "The Groaci don't know. They're a very cultured, gentle people. You can do irreparable harm to the reputation of Terrestrials if you insist—" "That's my decision," Retief said. "I have a job to do and we're wasting time." He crossed the room to his desk, opened a drawer and took out a slim-barreled needler. "This office is being watched. Not very efficiently, if I know the Groaci. I think I can get past them all right." "Where are you going with ... that?" Miss Meuhl stared at the needler. "What in the world—" "The Groaci won't waste any time destroying every piece of paper in their files relating to this thing. I have to get what I need before it's too late. If I wait for an official Inquiry Commission, they'll find nothing but blank smiles." "You're out of your mind!" Miss Meuhl stood up, quivering with indignation. "You're like a ... a...." "You and I are in a tight spot, Miss Meuhl. The logical next move for the Groaci is to dispose of both of us. We're the only ones who know what happened. Fith almost did the job this afternoon, but I bluffed him out—for the moment." Miss Meuhl emitted a shrill laugh. "Your fantasies are getting the better of you," she gasped. "In danger, indeed! Disposing of me! I've never heard anything so ridiculous." "Stay in this office. Close and safe-lock the door. You've got food and water in the dispenser. I suggest you stock up, before they shut the supply down. Don't let anyone in, on any pretext whatever. I'll keep in touch with you via hand-phone." "What are you planning to do?" "If I don't make it back here, transmit the sealed record of this afternoon's conversation, along with the information I've given you. Beam it through on a mayday priority. Then tell the Groaci what you've done and sit tight. I think you'll be all right. It won't be easy to blast in here and anyway, they won't make things worse by killing you. A force can be here in a week." "I'll do nothing of the sort! The Groaci are very fond of me! You ... Johnny-come-lately! Roughneck! Setting out to destroy—" "Blame it on me if it will make you feel any better," Retief said, "but don't be fool enough to trust them." He pulled on a cape, opened the door. "I'll be back in a couple of hours," he said. Miss Meuhl stared after him silently as he closed the door. It was an hour before dawn when Retief keyed the combination to the safe-lock and stepped into the darkened consular office. He looked tired. Miss Meuhl, dozing in a chair, awoke with a start. She looked at Retief, rose and snapped on a light, turned to stare. "What in the world—Where have you been? What's happened to your clothing?" "I got a little dirty. Don't worry about it." Retief went to his desk, opened a drawer and replaced the needler. "Where have you been?" Miss Meuhl demanded. "I stayed here—" "I'm glad you did," Retief said. "I hope you piled up a supply of food and water from the dispenser, too. We'll be holed up here for a week, at least." He jotted figures on a pad. "Warm up the official sender. I have a long transmission for Regional Headquarters." "Are you going to tell me where you've been?" "I have a message to get off first, Miss Meuhl," Retief said sharply. "I've been to the Foreign Ministry," he added. "I'll tell you all about it later." "At this hour? There's no one there...." "Exactly." Miss Meuhl gasped. "You mean you broke in? You burgled the Foreign Office?" "That's right," Retief said calmly. "Now—" "This is absolutely the end!" Miss Meuhl said. "Thank heaven I've already—" "Get that sender going, woman!" Retief snapped. "This is important." "I've already done so, Mr. Retief!" Miss Meuhl said harshly. "I've been waiting for you to come back here...." She turned to the communicator, flipped levers. The screen snapped aglow, and a wavering long-distance image appeared. "He's here now," Miss Meuhl said to the screen. She looked at Retief triumphantly. "That's good," Retief said. "I don't think the Groaci can knock us off the air, but—" "I have done my duty, Mr. Retief," Miss Meuhl said. "I made a full report to Regional Headquarters last night, as soon as you left this office. Any doubts I may have had as to the rightness of that decision have been completely dispelled by what you've just told me." Retief looked at her levelly. "You've been a busy girl, Miss Meuhl. Did you mention the six Terrestrials who were killed here?" "That had no bearing on the matter of your wild behavior! I must say, in all my years in the Corps, I've never encountered a personality less suited to diplomatic work." The screen crackled, the ten-second transmission lag having elapsed. "Mr. Retief," the face on the screen said, "I am Counsellor Pardy, DSO-1, Deputy Under-secretary for the region. I have received a report on your conduct which makes it mandatory for me to relieve you administratively, vice Miss Yolanda Meuhl, DAO-9. Pending the findings of a Board of Inquiry, you will—" Retief reached out and snapped off the communicator. The triumphant look faded from Miss Meuhl's face. "Why, what is the meaning—" "If I'd listened any longer, I might have heard something I couldn't ignore. I can't afford that, at this moment. Listen, Miss Meuhl," Retief went on earnestly, "I've found the missing cruiser." "You heard him relieve you!" "I heard him say he was going to, Miss Meuhl. But until I've heard and acknowledged a verbal order, it has no force. If I'm wrong, he'll get my resignation. If I'm right, that suspension would be embarrassing all around." "You're defying lawful authority! I'm in charge here now." Miss Meuhl stepped to the local communicator. "I'm going to report this terrible thing to the Groaci at once, and offer my profound—" "Don't touch that screen," Retief said. "You go sit in that corner where I can keep an eye on you. I'm going to make a sealed tape for transmission to Headquarters, along with a call for an armed task force. Then we'll settle down to wait." Retief ignored Miss Meuhl's fury as he spoke into the recorder. The local communicator chimed. Miss Meuhl jumped up, staring at it. "Go ahead," Retief said. "Answer it." A Groacian official appeared on the screen. "Yolanda Meuhl," he said without preamble, "for the Foreign Minister of the Groacian Autonomy, I herewith accredit you as Terrestrial Consul to Groac, in accordance with the advices transmitted to my government direct from the Terrestrial Headquarters. As consul, you are requested to make available for questioning Mr. J. Retief, former consul, in connection with the assault on two peace keepers and illegal entry into the offices of the Ministry for Foreign Affairs." "Why, why," Miss Meuhl stammered. "Yes, of course. And I do want to express my deepest regrets—" Retief rose, went to the communicator, assisted Miss Meuhl aside. "Listen carefully, Fith," he said. "Your bluff has been called. You don't come in and we don't come out. Your camouflage worked for nine years, but it's all over now. I suggest you keep your heads and resist the temptation to make matters worse than they are." "Miss Meuhl," Fith said, "a peace squad waits outside your consulate. It is clear you are in the hands of a dangerous lunatic. As always, the Groaci wish only friendship with the Terrestrials, but—" "Don't bother," Retief said. "You know what was in those files I looked over this morning." Retief turned at a sound behind him. Miss Meuhl was at the door, reaching for the safe-lock release.... "Don't!" Retief jumped—too late. The door burst inward. A crowd of crested Groaci pressed into the room, pushed Miss Meuhl back, aimed scatter guns at Retief. Police Chief Shluh pushed forward. "Attempt no violence, Terrestrial," he said. "I cannot promise to restrain my men." "You're violating Terrestrial territory, Shluh," Retief said steadily. "I suggest you move back out the same way you came in." "I invited them here," Miss Meuhl spoke up. "They are here at my express wish." "Are they? Are you sure you meant to go this far, Miss Meuhl? A squad of armed Groaci in the consulate?" "You are the consul, Miss Yolanda Meuhl," Shluh said. "Would it not be best if we removed this deranged person to a place of safety?" "You're making a serious mistake, Shluh," Retief said. "Yes," Miss Meuhl said. "You're quite right, Mr. Shluh. Please escort Mr. Retief to his quarters in this building—" "I don't advise you to violate my diplomatic immunity, Fith," Retief said. "As chief of mission," Miss Meuhl said quickly, "I hereby waive immunity in the case of Mr. Retief." Shluh produced a hand recorder. "Kindly repeat your statement, Madam, officially," he said. "I wish no question to arise later." "Don't be a fool, woman," Retief said. "Don't you see what you're letting yourself in for? This would be a hell of a good time for you to figure out whose side you're on." "I'm on the side of common decency!" "You've been taken in. These people are concealing—" "You think all women are fools, don't you, Mr. Retief?" She turned to the police chief and spoke into the microphone he held up. "That's an illegal waiver," Retief said. "I'm consul here, whatever rumors you've heard. This thing's coming out into the open, whatever you do. Don't add violation of the Consulate to the list of Groacian atrocities." "Take the man," Shluh said. | D. The Terrific |
Which baselines did they compare? | ### Introduction
Ever since the LIME algorithm BIBREF0 , "explanation" techniques focusing on finding the importance of input features in regard of a specific prediction have soared and we now have many ways of finding saliency maps (also called heat-maps because of the way we like to visualize them). We are interested in this paper by the use of such a technique in an extreme task that highlights questions about the validity and evaluation of the approach. We would like to first set the vocabulary we will use. We agree that saliency maps are not explanations in themselves and that they are more similar to attribution, which is only one part of the human explanation process BIBREF1 . We will prefer to call this importance mapping of the input an attribution rather than an explanation. We will talk about the importance of the input relevance score in regard to the model's computation and not make allusion to any human understanding of the model as a result. There exist multiple ways to generate saliency maps over the input for non-linear classifiers BIBREF2 , BIBREF3 , BIBREF4 . We refer the reader to BIBREF5 for a survey of explainable AI in general. We use in this paper Layer-Wise Relevance Propagation (LRP) BIBREF2 which aims at redistributing the value of the classifying function on the input to obtain the importance attribution. It was first created to “explain" the classification of neural networks on image recognition tasks. It was later successfully applied to text using convolutional neural networks (CNN) BIBREF6 and then Long-Short Term Memory (LSTM) networks for sentiment analysis BIBREF7 . Our goal in this paper is to test the limits of the use of such a technique for more complex tasks, where the notion of input importance might not be as simple as in topic classification or sentiment analysis. We changed from a classification task to a generative task and chose a more complex one than text translation (in which we can easily find a word to word correspondence/importance between input and output). We chose text summarization. We consider abstractive and informative text summarization, meaning that we write a summary “in our own words" and retain the important information of the original text. We refer the reader to BIBREF8 for more details on the task and the different variants that exist. Since the success of deep sequence-to-sequence models for text translation BIBREF9 , the same approaches have been applied to text summarization tasks BIBREF10 , BIBREF11 , BIBREF12 which use architectures on which we can apply LRP. We obtain one saliency map for each word in the generated summaries, supposed to represent the use of the input features for each element of the output sequence. We observe that all the saliency maps for a text are nearly identical and decorrelated with the attention distribution. We propose a way to check their validity by creating what could be seen as a counterfactual experiment from a synthesis of the saliency maps, using the same technique as in Arras et al. Arras2017. We show that in some but not all cases they help identify the important input features and that we need to rigorously check importance attributions before trusting them, regardless of whether or not the mapping “makes sense" to us. We finally argue that in the process of identifying the important input features, verifying the saliency maps is as important as the generation step, if not more. ### The Task and the Model
We present in this section the baseline model from See et al. See2017 trained on the CNN/Daily Mail dataset. We reproduce the results from See et al. See2017 to then apply LRP on it. ### Dataset and Training Task
The CNN/Daily mail dataset BIBREF12 is a text summarization dataset adapted from the Deepmind question-answering dataset BIBREF13 . It contains around three hundred thousand news articles coupled with summaries of about three sentences. These summaries are in fact “highlights" of the articles provided by the media themselves. Articles have an average length of 780 words and the summaries of 50 words. We had 287 000 training pairs and 11 500 test pairs. Similarly to See et al. See2017, we limit during training and prediction the input text to 400 words and generate summaries of 200 words. We pad the shorter texts using an UNKNOWN token and truncate the longer texts. We embed the texts and summaries using a vocabulary of size 50 000, thus recreating the same parameters as See et al. See2017. ### The Model
The baseline model is a deep sequence-to-sequence encoder/decoder model with attention. The encoder is a bidirectional Long-Short Term Memory(LSTM) cell BIBREF14 and the decoder a single LSTM cell with attention mechanism. The attention mechanism is computed as in BIBREF9 and we use a greedy search for decoding. We train end-to-end including the words embeddings. The embedding size used is of 128 and the hidden state size of the LSTM cells is of 254. ### Obtained Summaries
We train the 21 350 992 parameters of the network for about 60 epochs until we achieve results that are qualitatively equivalent to the results of See et al. See2017. We obtain summaries that are broadly relevant to the text but do not match the target summaries very well. We observe the same problems such as wrong reproduction of factual details, replacing rare words with more common alternatives or repeating non-sense after the third sentence. We can see in Figure 1 an example of summary obtained compared to the target one. The “summaries" we generate are far from being valid summaries of the information in the texts but are sufficient to look at the attribution that LRP will give us. They pick up the general subject of the original text. ### Layer-Wise Relevance Propagation
We present in this section the Layer-Wise Relevance Propagation (LRP) BIBREF2 technique that we used to attribute importance to the input features, together with how we adapted it to our model and how we generated the saliency maps. LRP redistributes the output of the model from the output layer to the input by transmitting information backwards through the layers. We call this propagated backwards importance the relevance. LRP has the particularity to attribute negative and positive relevance: a positive relevance is supposed to represent evidence that led to the classifier's result while negative relevance represents evidence that participated negatively in the prediction. ### Mathematical Description
We initialize the relevance of the output layer to the value of the predicted class before softmax and we then describe locally the propagation backwards of the relevance from layer to layer. For normal neural network layers we use the form of LRP with epsilon stabilizer BIBREF2 . We write down $R_{i\leftarrow j}^{(l, l+1)}$ the relevance received by the neuron $i$ of layer $l$ from the neuron $j$ of layer $l+1$ : $$\begin{split}
R_{i\leftarrow j}^{(l, l+1)} &= \dfrac{w_{i\rightarrow j}^{l,l+1}\textbf {z}^l_i + \dfrac{\epsilon \textrm { sign}(\textbf {z}^{l+1}_j) + \textbf {b}^{l+1}_j}{D_l}}{\textbf {z}^{l+1}_j + \epsilon * \textrm { sign}(\textbf {z}^{l+1}_j)} * R_j^{l+1} \\
\end{split}$$ (Eq. 7) where $w_{i\rightarrow j}^{l,l+1}$ is the network's weight parameter set during training, $\textbf {b}^{l+1}_j$ is the bias for neuron $j$ of layer $l+1$ , $\textbf {z}^{l}_i$ is the activation of neuron $i$ on layer $l$ , $\epsilon $ is the stabilizing term set to 0.00001 and $D_l$ is the dimension of the $l$ -th layer. The relevance of a neuron is then computed as the sum of the relevance he received from the above layer(s). For LSTM cells we use the method from Arras et al.Arras2017 to solve the problem posed by the element-wise multiplications of vectors. Arras et al. noted that when such computation happened inside an LSTM cell, it always involved a “gate" vector and another vector containing information. The gate vector containing only value between 0 and 1 is essentially filtering the second vector to allow the passing of “relevant" information. Considering this, when we propagate relevance through an element-wise multiplication operation, we give all the upper-layer's relevance to the “information" vector and none to the “gate" vector. ### Generation of the Saliency Maps
We use the same method to transmit relevance through the attention mechanism back to the encoder because Bahdanau's attention BIBREF9 uses element-wise multiplications as well. We depict in Figure 2 the transmission end-to-end from the output layer to the input through the decoder, attention mechanism and then the bidirectional encoder. We then sum up the relevance on the word embedding to get the token's relevance as Arras et al. Arras2017. The way we generate saliency maps differs a bit from the usual context in which LRP is used as we essentially don't have one classification, but 200 (one for each word in the summary). We generate a relevance attribution for the 50 first words of the generated summary as after this point they often repeat themselves. This means that for each text we obtain 50 different saliency maps, each one supposed to represent the relevance of the input for a specific generated word in the summary. ### Experimental results
In this section, we present our results from extracting attributions from the sequence-to-sequence model trained for abstractive text summarization. We first have to discuss the difference between the 50 different saliency maps we obtain and then we propose a protocol to validate the mappings. ### First Observations
The first observation that is made is that for one text, the 50 saliency maps are almost identical. Indeed each mapping highlights mainly the same input words with only slight variations of importance. We can see in Figure 3 an example of two nearly identical attributions for two distant and unrelated words of the summary. The saliency map generated using LRP is also uncorrelated with the attention distribution that participated in the generation of the output word. The attention distribution changes drastically between the words in the generated summary while not impacting significantly the attribution over the input text. We deleted in an experiment the relevance propagated through the attention mechanism to the encoder and didn't observe much changes in the saliency map. It can be seen as evidence that using the attention distribution as an “explanation" of the prediction can be misleading. It is not the only information received by the decoder and the importance it “allocates" to this attention state might be very low. What seems to happen in this application is that most of the information used is transmitted from the encoder to the decoder and the attention mechanism at each decoding step just changes marginally how it is used. Quantifying the difference between attention distribution and saliency map across multiple tasks is a possible future work. The second observation we can make is that the saliency map doesn't seem to highlight the right things in the input for the summary it generates. The saliency maps on Figure 3 correspond to the summary from Figure 1 , and we don't see the word “video" highlighted in the input text, which seems to be important for the output. This allows us to question how good the saliency maps are in the sense that we question how well they actually represent the network's use of the input features. We will call that truthfulness of the attribution in regard to the computation, meaning that an attribution is truthful in regard to the computation if it actually highlights the important input features that the network attended to during prediction. We proceed to measure the truthfulness of the attributions by validating them quantitatively. ### Validating the Attributions
We propose to validate the saliency maps in a similar way as Arras et al. Arras2017 by incrementally deleting “important" words from the input text and observe the change in the resulting generated summaries. We first define what “important" (and “unimportant") input words mean across the 50 saliency maps per texts. Relevance transmitted by LRP being positive or negative, we average the absolute value of the relevance across the saliency maps to obtain one ranking of the most “relevant" words. The idea is that input words with negative relevance have an impact on the resulting generated word, even if it is not participating positively, while a word with a relevance close to zero should not be important at all. We did however also try with different methods, like averaging the raw relevance or averaging a scaled absolute value where negative relevance is scaled down by a constant factor. The absolute value average seemed to deliver the best results. We delete incrementally the important words (words with the highest average) in the input and compared it to the control experiment that consists of deleting the least important word and compare the degradation of the resulting summaries. We obtain mitigated results: for some texts, we observe a quick degradation when deleting important words which are not observed when deleting unimportant words (see Figure 4 ), but for other test examples we don't observe a significant difference between the two settings (see Figure 5 ). One might argue that the second summary in Figure 5 is better than the first one as it makes better sentences but as the model generates inaccurate summaries, we do not wish to make such a statement. This however allows us to say that the attribution generated for the text at the origin of the summaries in Figure 4 are truthful in regard to the network's computation and we may use it for further studies of the example, whereas for the text at the origin of Figure 5 we shouldn't draw any further conclusions from the attribution generated. One interesting point is that one saliency map didn't look “better" than the other, meaning that there is no apparent way of determining their truthfulness in regard of the computation without doing a quantitative validation. This brings us to believe that even in simpler tasks, the saliency maps might make sense to us (for example highlighting the animal in an image classification task), without actually representing what the network really attended too, or in what way. We defined without saying it the counterfactual case in our experiment: “Would the important words in the input be deleted, we would have a different summary". Such counterfactuals are however more difficult to define for image classification for example, where it could be applying a mask over an image, or just filtering a colour or a pattern. We believe that defining a counterfactual and testing it allows us to measure and evaluate the truthfulness of the attributions and thus weight how much we can trust them. ### Conclusion
In this work, we have implemented and applied LRP to a sequence-to-sequence model trained on a more complex task than usual: text summarization. We used previous work to solve the difficulties posed by LRP in LSTM cells and adapted the same technique for Bahdanau et al. Bahdanau2014 attention mechanism. We observed a peculiar behaviour of the saliency maps for the words in the output summary: they are almost all identical and seem uncorrelated with the attention distribution. We then proceeded to validate our attributions by averaging the absolute value of the relevance across the saliency maps. We obtain a ranking of the word from the most important to the least important and proceeded to delete one or another. We showed that in some cases the saliency maps are truthful to the network's computation, meaning that they do highlight the input features that the network focused on. But we also showed that in some cases the saliency maps seem to not capture the important input features. This brought us to discuss the fact that these attributions are not sufficient by themselves, and that we need to define the counter-factual case and test it to measure how truthful the saliency maps are. Future work would look into the saliency maps generated by applying LRP to pointer-generator networks and compare to our current results as well as mathematically justifying the average that we did when validating our saliency maps. Some additional work is also needed on the validation of the saliency maps with counterfactual tests. The exploitation and evaluation of saliency map are a very important step and should not be overlooked. Figure 2: Representation of the propagation of the relevance from the output to the input. It passes through the decoder and attention mechanism for each previous decoding time-step, then is passed onto the encoder which takes into account the relevance transiting in both direction due to the bidirectional nature of the encoding LSTM cell. Figure 3: Left : Saliency map over the truncated input text for the second generated word “the”. Right : Saliency map over the truncated input text for the 25th generated word “investigation”. We see that the difference between the mappings is marginal. Figure 4: Summary from Figure 1 generated after deleting important and unimportant words from the input text. We observe a significant difference in summary degradation between the two experiments, where the decoder just repeats the UNKNOWN token over and over. | The baseline model is a deep sequence-to-sequence encoder/decoder model with attention. The encoder is a bidirectional Long-Short Term Memory(LSTM) cell BIBREF14 and the decoder a single LSTM cell with attention mechanism. The attention mechanism is computed as in BIBREF9 and we use a greedy search for decoding. We train end-to-end including the words embeddings. The embedding size used is of 128 and the hidden state size of the LSTM cells is of 254. |
Why did the Lieutenant go into labor early?
A. A slight depressurization in the space station shocked her body into labor.
B. Major Banes induced labor early because the baby was unusually large.
C. The stress of living in outer space caused her body to go into pre-term labor.
D. An asteroid crashed into the space station causing it to jerk unexpectedly. The Lieutenant fell and her water broke.
| SPATIAL DELIVERY BY RANDALL GARRETT Women on space station assignments shouldn't get pregnant. But there's a first time for everything. Here's the story of such a time——and an historic situation. [Transcriber's Note: This etext was produced from Worlds of If Science Fiction, October 1954. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed.] One thousand seventy-five miles above the wrinkled surface of Earth, a woman was in pain. There, high in the emptiness of space, Space Station One swung in its orbit. Once every two hours, the artificial satellite looped completely around the planet, watching what went on below. Outside its bright steel hull was the silence of the interplanetary vacuum; inside, in the hospital ward, Lieutenant Alice Britton clutched at the sheets of her bed in pain, then relaxed as it faded away. Major Banes looked at her and smiled a little. "How do you feel, Lieutenant?" She smiled back; she knew the pain wouldn't return for a few minutes yet. "Fine, doctor. It's no worse than I was expecting. How long will it before we can contact White Sands?" The major looked nervously at his wristwatch. "Nearly an hour. You'll be all right." "Certainly," she agreed, running a hand through her brown hair, "I'll be okay. Just you be on tap when I call." The major's grin broadened. "You don't think I'd miss a historical event like this, do you? You take it easy. We're over Eastern Europe now, but as soon as we get within radio range of New Mexico, I'll beam a call in." He paused, then repeated, "You just take it easy. Call the nurse if anything happens." Then he turned and walked out of the room. Alice Britton closed her eyes. Major Banes was all smiles and cheer now, but he hadn't been that way five months ago. She chuckled softly to herself as she thought of his blistering speech. "Lieutenant Britton, you're either careless or brainless; I don't know which! Your husband may be the finest rocket jockey in the Space Service, but that doesn't give him the right to come blasting up here on a supply rocket just to get you pregnant!" Alice had said: "I'm sure the thought never entered his mind, doctor. I know it never entered mine." "But that was two and a half months ago! Why didn't you come to me before this? Of all the tom-fool—" His voice had died off in suppressed anger. "I didn't know," she had said stolidly. "You know my medical record." "I know. I know." A puzzled frown had come over his face then, a frown which almost hid the green eyes that contrasted so startlingly with the flaming red of his hair. "The question is: what do we do next? We're not equipped for obstetrics up here." "Send me back down to Earth, of course." And he had looked up at her scathingly. "Lieutenant Britton, it is my personal opinion that you need your head examined, and not by a general practitioner, either! Why, I wouldn't let you get into an airplane, much less land on Earth in a rocket! If you think I'd permit you to subject yourself to eight gravities of acceleration in a rocket landing, you're daffy!" She hadn't thought of it before, but the major was right. The terrible pressure of a rocket landing would increase her effective body weight to nearly half a ton; an adult human being couldn't take that sort of punishment for long, much less the tiny life that was growing within her. So she had stayed on in the Space Station, doing her job as always. As Chief Radar Technician, she was important in the operation of the station. Her pregnancy had never made her uncomfortable; the slow rotation of the wheel-shaped station about its axis gave an effective gravity at the rim only half that of Earth's surface, and the closer to the hub she went, the less her weight became. According to the major, the baby was due sometime around the first of September. "Two hundred and eighty days," he had said. "Luckily, we can pinpoint it almost exactly. And at a maximum of half of Earth gravity, you shouldn't weigh more than seventy pounds then. You're to report to me at least once a week, Lieutenant." As the words went through her mind, another spasm of pain hit her, and she clenched her fists tightly on the sheets again. It went away, and she took a deep breath. Everything had been fine until today. And then, only half an hour ago, a meteor had hit the radar room. It had been only a tiny bit of rock, no bigger than a twenty-two bullet, and it hadn't been traveling more than ten miles per second, but it had managed to punch its way through the shielding of the station. The self-sealing walls had closed the tiny hole quickly, but even in that short time, a lot of air had gone whistling out into the vacuum of space. The depressurization hadn't hurt her too much, but the shock had been enough to start labor. The baby was going to come two months early. She relaxed a little more, waiting for the next pain. There was nothing to worry about; she had absolute faith in the red-haired major. The major himself was not so sure. He sat in his office, massaging his fingertips and looking worriedly at the clock on the wall. The Chief Nurse at a nearby desk took off her glasses and looked at him speculatively. "Something wrong, doctor?" "Incubator," he said, without taking his eyes off the clock. "I beg your pardon?" "Incubator. We can't deliver a seven-month preemie without an incubator." The nurse's eyes widened. "Good Lord! I never thought of that! What are you going to do?" "Right now, I can't do anything. I can't beam a radio message through to the Earth. But as soon as we get within radio range of White Sands, I'll ask them to send up an emergency rocket with an incubator. But—" "But what?" "Will we have time? The pains are coming pretty fast now. It will be at least three hours before they can get a ship up here. If they miss us on the next time around, it'll be five hours. She can't hold out that long." The Chief Nurse turned her eyes to the slowly moving second hand of the wall clock. She could feel a lump in her throat. Major Banes was in the Communications Center a full five minutes before the coastline of California appeared on the curved horizon of the globe beneath them. He had spent the hour typing out a complete report of what had happened to Alice Britton and a list of what he needed. He handed it to the teletype operator and paced the floor impatiently as he waited for the answer. When the receiver teletype began clacking softly, he leaned over the page, waiting anxiously for every word. WHITE SANDS ROCKET BASE 4 JULY 1984 0913 HRS URGENT TO: MAJ PETER BANES (MC) 0-266118 SS-1 MEDICAL OFFICER FROM: GEN DAVID BARRETT 0-199515 COMMANDING WSRB ROCKET. ORBIT NOW BEING COMPUTED FOR RENDEZVOUS WITH SS-1 AS OF NEXT PASSAGE ABOVE USA. CAPT. JAMES BRITTON PILOTING. MEDICS LOADING SHIP TWELVE WITH INCUBATOR AND OTHER SUPPLIES. BASE OBSTETRICIAN LT COL GATES ALSO COMING TO ASSIST IN DELIVERY. HANG ON. OVER. Banes nodded and turned to the operator. "I want a direct open telephone line to my office in case I have to get another message to the base before we get out of range again." He turned and left through the heavy door. Each room of the space station was protected by airtight doors and individual heating units; if some accident, such as a really large meteor hit, should release the air from one room, nearby rooms would be safe. Banes' next stop was the hospital ward. Alice Britton was resting quietly, but there were lines of strain around her eyes which hadn't been there an hour before. "How's it coming, Lieutenant?" She smiled, but another spasm hit her before she could answer. After a time, she said: "I'm doing fine, but you look as if you'd been through the mill. What's eating you?" He forced a nervous smile. "Nothing but the responsibility. You're going to be a very famous woman, you know. You'll be the mother of the first child born in space. And it's my job to see to it that you're both all right." She grinned. "Another Dr. Dafoe?" "Something on that order, I suppose. But it won't be all my glory. Colonel Gates, the O.B. man, was supposed to come up for the delivery in September, so when White Sands contacted us, they said he was coming immediately." He paused, and a genuine smile crossed his face. "Your husband is bringing him up." "Jim! Coming up here? Wonderful! But I'm afraid the colonel will be too late. This isn't going to last that long." Banes had to fight hard to keep his face smiling when she said that, but he managed an easy nod. "We'll see. Don't hurry it, though. Let nature take its course. I'm not such a glory hog that I'd not let Gates have part of it—or all of it, for that matter. Relax and take it easy." He went on talking, trying to keep the conversation light, but his eyes kept wandering to his wristwatch, timing Alice's pain intervals. They were coming too close together to suit him. There was a faint rap, and the heavy airtight door swung open to admit the Chief Nurse. "There's a message for you in your office, doctor. I'll send a nurse in to be with her." He nodded, then turned back to Alice. "Stiff uppah lip, and all that sort of rot," he said in a phony British accent. "Oh, raw ther , old chap," she grinned. Back in his office, Banes picked up the teletype flimsy. WHITE SANDS ROCKET BASE 4 JULY 1984 0928 HRS URGENT TO: MAJ PETER BANES (MC) 0-266118 SS-1 MEDICAL OFFICER FROM: GEN DAVID BARRETT 0-199515 COMMANDING WSRB ROCKET. ORBIT COMPUTED FOR RENDEZVOUS AT 1134 HRS MST. CAPT BRITTON SENDS PERSONAL TO LT BRITTON AS FOLLOWS: HOLD THE FORT, BABY, THE WHOLE WORLD IS PRAYING FOR YOU. OUT. Banes sat on the edge of his desk, pounding a fist into the palm of his left hand. "Two hours. It isn't soon enough. She'll never hold out that long. And we don't have an incubator." His voice was a clipped monotone, timed with the rhythmic slamming of his fist. The Chief Nurse said: "Can't we build something that will do until the rocket gets here?" Banes looked at her, his face expressionless. "What would we build it out of? There's not a spare piece of equipment in the station. It costs money to ship material up here, you know. Anything not essential is left on the ground." The phone rang. Banes picked it up and identified himself. The voice at the other end said: "This is Communications, Major. I tape recorded all the monitor pickups from the Earth radio stations, and it looks as though the Space Service has released the information to the public. Lieutenant Britton's husband was right when he said the whole world's praying for her. Do you want to hear the tapes?" "Not now, but thanks for the information." He hung up and looked into the Chief Nurse's eyes. "They've released the news to the public." She frowned. "That really puts you on the spot. If the baby dies, they'll blame you." Banes slammed his fist to the desk. "Do you think I give a tinker's dam about that? I'm interested in saving a life, not in worrying about what people may think!" "Yes, sir. I just thought—" "Well, think about something useful! Think about how we're going to save that baby!" He paused as he saw her eyes. "I'm sorry, Lieutenant. My nerves are all raw, I guess. But, dammit, my field is space medicine. I can handle depressurization, space sickness, and things like that, but I don't know anything about babies! I know what I read in medical school, and I watched a delivery once, but that's all I know. I don't even have any references up here; people aren't supposed to go around having babies on a space station!" "It's all right, doctor. Shall I prepare the delivery room?" His laugh was hard and short. "Delivery room! I wish to Heaven we had one! Prepare the ward room next to the one she's in now, I guess. It's the best we have. "So help me Hannah, I'm going to see some changes made in regulations! A situation like this won't happen again!" The nurse left quietly. She knew Banes wasn't really angry at the Brittons; it was simply his way of letting off steam to ease the tension within him. The slow, monotonous rotation of the second hand on the wall clock seemed to drag time grudgingly along with it. Banes wished he could smoke to calm his raw nerves, but it was strictly against regulations. Air was too precious to be used up by smoking. Every bit of air on board had had to be carried up in rockets when the station was built in space. The air purifiers in the hydroponics section could keep the air fresh enough for breathing, but fire of any kind would overtax the system, leaving too little oxygen in the atmosphere. It was a few minutes of ten when he decided he'd better get back to Alice Britton. She was trying to read a book between spasms, but she wasn't getting much read. She dropped it to the floor when he came in. "Am I glad to see you! It won't be long now." She looked at him analytically. "Say! Just what is eating you? You look more haggard than I do!" Again he tried to force a smile, but it didn't come off too well. "Nothing serious. I just want to make sure everything comes out all right." She smiled. "It will. You're all set. You ordered the instruments months ago. Or did you forget something?" That hit home, but he just grinned feebly. "I forgot to get somebody to boil water." "Whatever for?" "Coffee, of course. Didn't you know that? Papa always heats up the water; that keeps him out of the way, and the doctor has coffee afterwards." Alice's hands grasped the sheet again, and Banes glanced at his watch. Ninety seconds! It was long and hard. When the pain had ebbed away, he said: "We've got the delivery room all ready. It won't be much longer now." "I'll say it won't! How about the incubator?" There was a long pause. Finally, he said softly: "There isn't any incubator. I didn't take the possibility of a premature delivery into account. It's my fault. I've done what I could, though; the ship is bringing one up. I—I think we'll be able to keep the child alive until—" He stopped. Alice was bubbling up with laughter. "Lieutenant! Lieutenant Britton! Alice! This is no time to get hysterical! Stop it!" Her laughter slowed to a chuckle. " Me get hysterical! That's a good one! What about you? You're so nervous you couldn't sip water out of a bathtub without spilling it!" He blinked. "What do you mean?" Another pain came, and he had to wait until it was over before he got her answer. "Doctor," she said, "I thought you would have figured it out. Ask yourself just one question. Ask yourself, 'Why is a space station like an incubator?'" Space Ship Twelve docked at Space Station One at exactly eleven thirty-four, and two men in spacesuits pushed a large, bulky package through the airlock. Major Peter Banes, haggard but smiling, met Captain Britton in the corridor as he and the colonel entered the hospital ward. Banes nodded to Colonel Gates, then turned to Britton. "I don't know whether to congratulate you or take a poke at you, Captain, but I suppose congratulations come first. Your son, James Edward Britton II, is doing fine, thank you." "You mean— already ?" The colonel said nothing, but he raised an eyebrow. "Over an hour ago," said Banes. "But—but—the incubator—" Banes' grin widened. "We'll put the baby in it, now that we've got it, but it really isn't necessary. Your wife figured that one out. A space station is a kind of incubator itself, you see. It protects us poor, weak humans from the terrible conditions of space. So all we had to do was close up one of the airtight rooms, sterilize it, warm it up, and put in extra oxygen from the emergency tanks. Young James is perfectly comfortable." "Excellent, Major!" said the colonel. "Don't thank me. It was Captain Britton's wife who—" But Captain Britton wasn't listening any more. He was headed toward his wife's room at top speed. | A. A slight depressurization in the space station shocked her body into labor. |
What can you best infer about the connection the main character had with his wife?
A. He had a connection with her worth more than what could be captured in an object.
B. He had a connection with her that would end once he left for war.
C. His connection with her was not strong enough to withstand time.
D. His connection with her was not as strong as he thought.
| HOMECOMING BY MIGUEL HIDALGO What lasts forever? Does love? Does death?... Nothing lasts forever.... Not even forever [Transcriber's Note: This etext was produced from Worlds of If Science Fiction, April 1958. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed.] The large horse plodded slowly over the shifting sand. The rider was of medium size, with huge, strong hands and seemingly hollow eyes. Strange eyes, alive and aflame. They had no place in the dust-caked, tired body, yet there they were, seeking, always seeking—searching the clear horizon, and never seeming to find what they sought. The horse moved faster now. They were nearing a river; the water would be welcome on tired bodies and dry throats. He spurred his horse, and when they reached the water's edge, he dismounted and unsaddled the horse. Then both man and horse plunged headlong into the waiting torrent, deep into the cool embrace of the clear liquid. They soaked it into their pores and drank deeply of it, feeling life going once more through their veins. Satisfied, they lifted themselves from the water, and the man lay down on the yellow sand of the river bank to sleep. When he awoke, the sun was almost setting. The bright shafts of red light spilled across the sky, making the mountains silent scarlet shadows on the face of the rippling water. Quickly he gathered driftwood, and built a small fire. From his pack he removed some of the coffee he had found in one of the ruined cities. He brought water from the river in the battered coffee-pot he had salvaged, and while he waited for it to boil, he went to his horse, Conqueror, stroking his mane and whispering in his ear. Then he led him silently to a grassy slope where he hobbled him and left him for the night. In the fading light, he ate the hard beef jerky and drank the scalding coffee. Refreshed and momentarily content, he sat staring into the dying fire, seeing the bright glowing coals as living fingers clutching at the wood in consuming embrace, taking all and returning nothing but ashes. Slowly his eyelids yielded. His body sagged, and blood seemed to fill his brain, bathing it in a gentle, warm flood. He slept. His brain slept. But the portion of his brain called memory stirred. It was all alone; all else was at rest. Images began to appear, drawn from inexhaustible files, wherein are kept all thoughts, past, present, and future.... It was the night before he was to go overseas. World War III had been declared, and he had enlisted, receiving his old rank of captain. He was with his wife in the living room of their home. They had put the children to bed—their sons—and now sat on the couch, watching the blazing fire. It was then that he had showed it to her. "I've got something to tell you, and something to show you." He had removed the box from his pocket and opened it. And heard her cry of surprised joy. "Oh, a ring, and it's a diamond, too!" she cried in her rich, happy voice which always seemed to send a thrill through his body. "It's for you; so long as you wear it, I'll come back, even from the dead, if need be. Read the inscription." She held the ring up to the light and read aloud, "It is forever." Then she had slipped the ring on her finger and her arms around him. He held her very close, feeling the warmth from her body flowing into his and making him oblivious to everything except that she was there in his arms and that he was sinking deep, deep into a familiar sea, where he had been many times before but each time found something new and unexplored, some vastly different emotion he could never quite explain. "Wait!" she cried. "I've something for you, too." She took off the locket she wore about her neck and held it up to the shimmering light, letting it spin at the end of its chain. It caught the shadows of the fire and reflected them, greatly magnified, over the room. It was in the shape of a star, encrusted with emeralds, with one large ruby in the center. When he opened it, he found a picture of her in one side, and in the other a picture of the children. He took her in his arms again, and loosened her long, black hair, burying his face in it for a moment. Then he kissed her, and instantly was drawn down into the abyss which seemed to have no beginning or any end. The next morning had been bleak and gray. The mist clung to the wet, sodden ground, and the air was heavy in his lungs. He had driven off in the jeep the army had sent for him, watching her there on the porch until the mist swirled around her feet and she ran back into the house and slammed the door. His cold fingers found the locket, making a little bulge under his uniform, and the touch of it seemed to warm the blood in his veins. Three days later they had landed in Spain, merged with another division, then crossed the Pyrenees into France, and finally to Paris where the fighting had begun. Already the city was a silent graveyard, littered with the rubble of towers and cathedrals which had once been great. Three years later they were on the road to Moscow. Over a thousand miles lay behind, a dead man on every foot of those miles. Yet victory was near. The Russians had not yet used the H-bomb; the threat of annihilation by the retaliation forces had been too great. He had done well in the war, and had been decorated many times for bravery in action. Now he felt the victory that seemed to be in the air, and he had wished it would come quickly, so that he might return to her. Home. The very feel of the word was everything a battle-weary soldier needed to make him fight harder and live longer. Suddenly he had become aware of a droning, wooshing sound above him. It grew louder and louder until he knew what it was. "Heavy bombers!" The alarm had sounded, and the men had headed for their foxholes. But the planes had passed over, the sun glinting on their bellies, reflecting a blinding light. They were bound for bigger, more important targets. When the all-clear had sounded, the men clambered from their shelters. An icy wind swept the field, bringing with it clouds which covered the sun. A strange fear had gripped him then.... Across the Atlantic, over the pole, via Alaska, the great bombers flew. In cities, great and small, the air raid sirens sounded, high screaming noises which had jarred the people from sleep in time to die. The defending planes roared into the sky to intercept the on-rushing bombers. The horrendous battle split the universe. Many bombers fell, victims of fanatical suicide planes, or of missiles that streaked across the sky which none could escape. But too many bombers got through, dropping their deadly cargo upon the helpless cities. And not all the prayers or entreaties to any God had stopped their carnage. First there had been the red flashes that melted buildings into molten streams, and then the great triple-mushroom cloud filled with the poisonous gases that the wind swept away to other cities, where men had not died quickly and mercifully, but had rotted away, leaving shreds of putrid flesh behind to mark the places where they had crawled. The retaliatory forces had roared away to bomb the Russian cities. Few, if any, had returned. Too much blood and life were on their hands. Those who had remained alive had found a resting place on the crown of some distant mountain. Others had preferred the silent peaceful sea, where flesh stayed not long on bones, and only darting fishes and merciful beams of filtered light found their aluminum coffins. The war had ended. To no avail. Neither side had won. Most of the cities and the majority of the population of both countries had been destroyed. Even their governments had vanished, leaving a silent nothingness. The armies that remained were without leaders, without sources of supplies, save what they could forage and beg from an unfriendly people. They were alone now, a group of tired, battered men, for whom life held nothing. Their families had long since died, their bodies turned to dust, their spirits fled on the winds to a new world. Yet these remnants of an army must return—or at least try. Their exodus was just beginning. Somehow he had managed to hold together the few men left from his force. He had always nourished the hope that she might still be alive. And now that the war was over he had to return—had to know whether she was still waiting for him. They had started the long trek. Throughout Europe anarchy reigned. He and his men were alone. All they could do now was fight. Finally they reached the seaport city of Calais. With what few men he had left, he had commandeered a small yacht, and they had taken to the sea. After months of storms and bad luck, they had been shipwrecked somewhere off the coast of Mexico. He had managed to swim ashore, and had been found by a fisherman's family. Many months he had spent swimming and fishing, recovering his strength, inquiring about the United States. The Mexicans had spoken with fear of the land across the Rio Grande. All its great cities had been destroyed, and those that had been only partially destroyed were devoid of people. The land across the Rio Grande had become a land of shadows. The winds were poisoned, and the few people who might have survived, were crazed and maimed by the blasts. Few men had dared cross the Rio Grande into "El Mundo gris de Noviembre"—the November world. Those who had, had never returned. In time he had traveled north until he reached the Rio Grande. He had waded into the muddy waters and somehow landed on the American side. In the November world. It was rightly called. The deserts were long. All plant life had died, leaving to those once great fertile stretches, nothing but the sad, temporal beauty that comes with death. No people had he seen. Only the ruins of what had once been their cities. He had walked through them, and all that he had seen were the small mutant rodents, and all that he had heard was the occasional swish of the wind as it whisked along what might have been dead leaves, but wasn't. He had been on the trail for a long time. His food was nearly exhausted. The mountains were just beginning, and he hoped to find food there. He had not found food, but his luck had been with him. He had found a horse. Not a normal horse, but a mutation. It was almost twice as large as a regular horse. Its skin seemed to shimmer and was like glassy steel to the touch. From the center of its forehead grew a horn, straight out, as the horn of a unicorn. But most startling of all were the animal's eyes which seemed to speak—a silent mental speech, which he could understand. The horse had looked up as he approached it and seemed to say: "Follow me." And he had followed. Over a mountain, until they came to a pass, and finally to a narrow path which led to an old cabin. He had found it empty, but there were cans of food and a rifle and many shells. He had remained there a long time—how long he could not tell, for he could only measure time by the cycles of the sun and the moon. Finally he had taken the horse, the rifle and what food was left, and once again started the long journey home. The farther north he went, the more life seemed to have survived. He had seen great herds of horses like his own, stampeding across the plains, and strange birds which he could not identify. Yet he had seen no human beings. But he knew he was closer now. Closer to home. He recognized the land. How, he did not know, for it was much changed. A sensing, perhaps, of what it had once been. He could not be more than two days' ride away. Once he was through this desert, he would find her, he would be with her once again; all would be well, and his long journey would be over. The images faded. Even memory slept in a flow of warm blood. Body and mind slept into the shadows of the dawn. He awoke and stretched the cramped muscles of his body. At the edge of the water he removed his clothes and stared at himself in the rippling mirror. His muscles were lean and hard, evenly placed throughout the length of his frame. A deep ridge ran down the length of his torso, separating the muscles, making the chest broad. Well satisfied with his body, he plunged into the cold water, deep down, until he thought his lungs would burst; then swiftly returned to the clean air, tingling in every pore. He dried himself and dressed. Conqueror was eating the long grass near the stream. Quickly he saddled him. No time for breakfast. He would ride all day and the next night. And he would be home. Still northward. The hours crawled slower than a dying man. The sun was a torch that pierced his skin, seeming to melt his bones into a burning stream within his body. But day at last gave way to night, and the sun to the moon. The torch became a white pock-marked goddess, with streaming hair called stars. In the moonlight he had not seen the crater until he was at its very edge. Even then he might not have seen it had not the horse stopped suddenly. The wind swirled through its vast emptiness, slapping his face with dusty hands. For a moment he thought he heard voices—mournful, murmuring voices, echoing up from the misty depths. He turned quickly away and did not look back. Night paled into day; day burned into night. There were clouds in the sky now, and a gentle wind caressed the sweat from his tired body. He stopped. There it was! Barely discernible through the moonlight, he saw it. Home. Quickly he dismounted and ran. Now he could see a small light in the window, and he knew they were there. His breath came in hard ragged gulps. At the window he peered in, and as his eyes became accustomed to the inner gloom, he saw how bare the room was. No matter. Now that he was home he would build new furniture, and the house would be even better than it had been before. Then he saw her. She was sitting motionless in a straight wooden chair beside the fireplace, the feeble light cast by the embers veiling her in mauve shadows. He waited, wondering if she were.... Presently she stirred like a restless child in sleep, then moved from the chair to the pile of wood near the hearth, and replenished the fire. The wood caught quickly, sending up long tongues of flame, and forming a bright pool of light around her. His blood froze. The creature illuminated by the firelight was a monster. Large greasy scales covered its face and arms, and there was no hair on its head. Its gums were toothless cavities in a sunken, mumbling mouth. The eyes, turned momentarily toward the window, were empty of life. "No, no!" he cried soundlessly. This was not his house. In his delirium he had only imagined he had found it. He had been searching so long. He would go on searching. He was turning wearily away from the window when the movement of the creature beside the fire held his attention. It had taken a ring from one skeleton-like finger and stood, turning the ring slowly as if trying to decipher some inscription inside it. He knew then. He had come home. Slowly he moved toward the door. A great weakness was upon him. His feet were stones, reluctant to leave the earth. His body was a weed, shriveled by thirst. He grasped the doorknob and clung to it, looking up at the night sky and trying to draw strength from the wind that passed over him. It was no use. There was no strength. Only fear—a kind of fear he had never known. He fumbled at his throat, his fingers crawling like cold worms around his neck until he found the locket and the clasp which had held it safely through endless nightmare days and nights. He slipped the clasp and the locket fell into his waiting hand. As one in a dream, he opened it, and stared at the pictures, now in the dim moonlight no longer faces of those he loved, but grey ghosts from the past. Even the ruby had lost its glow. What had once been living fire was now a dull glob of darkness. "Nothing is forever!" He thought he had shouted the words, but only a thin sound, the sound of leaves ruffled by the wind, came back to him. He closed the locket and fastened the clasp, and hung it on the doorknob. It moved slowly in the wind, back and forth, like a pendulum. "Forever—forever. Only death is forever." He could have sworn he heard the words. He ran. Away from the house. To the large horse with a horn in the center of its forehead, like a unicorn. Once in the saddle, the spurt of strength left him. His shoulders slumped, his head dropped onto his chest. Conqueror trotted away, the sound of his hooves echoing hollowly in the vast emptiness. | A. He had a connection with her worth more than what could be captured in an object. |
How is DIRL evaluated? | ### Introduction
Sentiment analysis aims to predict sentiment polarity of user-generated data with emotional orientation like movie reviews. The exponentially increase of online reviews makes it an interesting topic in research and industrial areas. However, reviews can span so many different domains and the collection and preprocessing of large amounts of data for new domains is often time-consuming and expensive. Therefore, cross-domain sentiment analysis is currently a hot topic, which aims to transfer knowledge from a label-rich source domain (S) to the label-few target domain (T). In recent years, one of the most popular frameworks for cross-domain sentiment analysis is the domain invariant representation learning (DIRL) framework BIBREF0, BIBREF1, BIBREF2, BIBREF3, BIBREF4. Methods of this framework follow the idea of extracting a domain-invariant feature representation, in which the data distributions of the source and target domains are similar. Based on the resultant representations, they learn the supervised classifier using source rich labeled data. The main difference among these methods is the applied technique to force the feature representations to be domain-invariant. However, in this work, we discover that applying DIRL may harm domain adaptation in the situation that the label distribution $\rm {P}(\rm {Y})$ shifts across domains. Specifically, let $\rm {X}$ and $\rm {Y}$ denote the input and label random variable, respectively, and $G(\rm {X})$ denote the feature representation of $\rm {X}$. We found out that when $\rm {P}(\rm {Y})$ changes across domains while $\rm {P}(\rm {X}|\rm {Y})$ stays the same, forcing $G(\rm {X})$ to be domain-invariant will make $G(\rm {X})$ uninformative to $\rm {Y}$. This will, in turn, harm the generation of the supervised classifier to the target domain. In addition, for the more general condition that both $\rm {P}(\rm {Y})$ and $\rm {P}(\rm {X}|\rm {Y})$ shift across domains, we deduced a conflict between the object of making the classification error small and that of making $G(\rm {X})$ domain-invariant. We argue that the problem is worthy of studying since the shift of $\rm {P}(\rm {Y})$ exists in many real-world cross-domain sentiment analysis tasks BIBREF0. For example, the marginal distribution of the sentiment of a product can be affected by the overall social environment and change in different time periods; and for different products, their marginal distributions of the sentiment are naturally considered different. Moreover, there are many factors, such as the original data distribution, data collection time, and data clearing method, that can affect $\rm {P}(\rm {Y})$ of the collected target domain unlabeled dataset. Note that in the real-world cross-domain tasks, we do not know the labels of the collected target domain data. Thus, we cannot previously align its label distribution $\rm {P}_T(\mathbf {Y})$ with that of source domain labeled data $\rm {P}_S(\mathbf {Y})$, as done in many previous works BIBREF0, BIBREF2, BIBREF5, BIBREF4, BIBREF6, BIBREF7. To address the problem of DIRL resulted from the shift of $\rm {P}(\rm {Y})$, we propose a modification to DIRL, obtaining a weighted domain-invariant representation learning (WDIRL) framework. This framework additionally introduces a class weight $\mathbf {w}$ to weigh source domain examples by class, hoping to make $\rm {P}(\rm {Y})$ of the weighted source domain close to that of the target domain. Based on $\mathbf {w}$, it resolves domain shift in two steps. In the first step, it forces the marginal distribution $\rm {P}(\rm {X})$ to be domain-invariant between the target domain and the weighted source domain instead of the original source, obtaining a supervised classifier $\rm {P}_S(\rm {Y}|\rm {X}; \mathbf {\Phi })$ and a class weight $\mathbf {w}$. In the second step, it resolves the shift of $\rm {P}(\rm {Y}|\rm {X})$ by adjusting $\rm {P}_S(\rm {Y}|\rm {X}; \mathbf {\Phi })$ using $\mathbf {w}$ for label prediction in the target domain. We detail these two steps in §SECREF4. Moreover, we will illustrate how to transfer existing DIRL models to their WDIRL counterparts, taking the representative metric-based CMD model BIBREF3 and the adversarial-learning-based DANN model BIBREF2 as an example, respectively. In summary, the contributions of this paper include: ($\mathbf {i}$) We theoretically and empirically analyse the problem of DIRL for domain adaptation when the marginal distribution $\rm {P}(\rm {Y})$ shifts across domains. ($\mathbf {ii}$) We proposed a novel method to address the problem and show how to incorporate it with existent DIRL models. ($\mathbf {iii}$) Experimental studies on extensive cross-domain sentiment analysis tasks show that models of our WDIRL framework can greatly outperform their DIRL counterparts. ### Preliminary and Related Work ::: Domain Adaptation
For expression consistency, in this work, we consider domain adaptation in the unsupervised setting (however, we argue that our analysis and solution also applies to the supervised and semi-supervised domain adaptation settings). In the unsupervised domain adaptation setting, there are two different distributions over $\rm {X} \times \rm {Y}$: the source domain $\rm {P}_S(\rm {X},\rm {Y})$ and the target domain $\rm {P}_T(\rm {X},\rm {Y})$. And there is a labeled data set $\mathcal {D}_S$ drawn $i.i.d$ from $\rm {P}_S(\rm {X},\rm {Y})$ and an unlabeled data set $\mathcal {D}_T$ drawn $i.i.d.$ from the marginal distribution $\rm {P}_T(\rm {X})$: The goal of domain adaptation is to build a classier $f:\rm {X} \rightarrow \rm {Y}$ that has good performance in the target domain using $\mathcal {D}_S$ and $\mathcal {D}_T$. For this purpose, many approaches have been proposed from different views, such as instance reweighting BIBREF8, pivot-based information passing BIBREF9, spectral feature alignment BIBREF10 subsampling BIBREF11, and of course the domain-invariant representation learning BIBREF12, BIBREF13, BIBREF14, BIBREF15, BIBREF16, BIBREF17, BIBREF18, BIBREF19, BIBREF20, BIBREF21, BIBREF22. ### Preliminary and Related Work ::: Domain Invariant Representation Learning
Domain invariant representation learning (DIRL) is a very popular framework for performing domain adaptation in the cross-domain sentiment analysis field BIBREF23, BIBREF4, BIBREF24, BIBREF7. It is heavily motivated by the following theorem BIBREF25. Theorem 1 For a hypothesis $h$, Here, $\mathcal {L}_S(h)$ denotes the expected loss with hypothesis $h$ in the source domain, $\mathcal {L}_T(h)$ denotes the counterpart in the target domain, $d_1$ is a measure of divergence between two distributions. Based on Theorem UNKREF3 and assuming that performing feature transform on $\rm {X}$ will not increase the values of the first and third terms of the right side of Ineq. (DISPLAY_FORM4), methods of the DIRL framework apply a feature map $G$ onto $\rm {X}$, hoping to obtain a feature representation $G(\rm {X})$ that has a lower value of ${d}_{1}(\rm {P}_S(G(\rm {X})), \rm {P}_T(G(\rm {X})))$. To this end, different methods have been proposed. These methods can be roughly divided into two directions. The first direction is to design a differentiable metric to explicitly evaluate the discrepancy between two distributions. We call methods of this direction as the metric-based DIRL methods. A representative work of this direction is the center-momentum-based model proposed by BIBREF3. In that work, they proposed a central moment discrepancy metric (CMD) to evaluate the discrepancy between two distributions. Specifically, let denote $\rm {X}_S$ and $\rm {X}_T$ an $M$ dimensional random vector on the compact interval $[a; b]^M$ over distribution $\rm {P}_S$ and $\rm {P}_T$, respectively. The CMD loss between $\rm {P}_S$ and $\rm {P}_T$ is defined by: Here, $\mathbb {E}(\rm {X})$ denotes the expectation of $\rm {X}$ over distribution $\rm {P}_S(\rm {X})$, and is the $k$-th momentum, where $\rm {X}_i$ denotes the $i^{th}$ dimensional variable of $\rm {X}$. The second direction is to perform adversarial training between the feature generator $G$ and a domain discriminator $D$. We call methods of this direction as the adversarial-learning-based methods. As a representative, BIBREF2 trained $D$ to distinguish the domain of a given example $x$ based on its representation $G(x)$. At the same time, they encouraged $G$ to deceive $D$, i.e., to make $D$ unable to distinguish the domain of $x$. More specifically, $D$ was trained to minimize the loss: over its trainable parameters, while in contrast $G$ was trained to maximize $\mathcal {L}_d$. According to the work of BIBREF26, this is equivalent to minimize the Jensen-shannon divergence BIBREF27, BIBREF28 $\text{JSD}(\rm {P}_S, \rm {P}_T)$ between $\rm {P}_S(G(\rm {X}))$ and $\rm {P}_T(G(\rm {X}))$ over $G$. Here, for a concise expression, we write $\rm {P}$ as the shorthand for $\rm {P}(G(\rm {X}))$. The task loss is the combination of the supervised learning loss $\mathcal {L}_{sup}$ and the domain-invariant learning loss $\mathcal {L}_{inv}$, which are defined on $\mathcal {D}_S$ only and on the combination of $\mathcal {D}_S$ and $\mathcal {D}_T$, respectively: Here, $\alpha $ is a hyper-parameter for loss balance, and the aforementioned domain adversarial loss $\text{JSD}(\rm {P}_S, \rm {P}_T)$ and $\text{CMD}_K$ are two concrete forms of $\mathcal {L}_{inv}$. ### Problem of Domain-Invariant Representation Learning
In this work, we found out that applying DIRL may harm domain adaptation in the situation that $\rm {P}(\rm {Y})$ shifts across domains. Specifically, when $\rm {P}_S(\rm {Y})$ differs from $\rm {P}_T(\rm {Y})$, forcing the feature representations $G(\rm {X})$ to be domain-invariant may increase the value of $\mathcal {L}_S(h)$ in Ineq. (DISPLAY_FORM4) and consequently increase the value of $\mathcal {L}_T(h)$, which means the decrease of target domain performance. In the following, we start our analysis under the condition that $\rm {P}_S(\rm {X}|\rm {Y})=\rm {P}_T(\rm {X}|\rm {Y})$. Then, we consider the more general condition that $\rm {P}_S(\rm {X}|\rm {Y})$ also differs from $\rm {P}_T(\rm {X}|\rm {Y})$. When $\rm {P}_S(\rm {X}|\rm {Y})=\rm {P}_T(\rm {X}|\rm {Y})$, we have the following theorem. Theorem 2 Given $\rm {P}_S(\rm {X}|\rm {Y})=\rm {P}_T(\rm {X}|\rm {Y})$, if $\rm {P}_S(\rm {Y}=i) \ne \rm {P}_T(\rm {Y}=i)$ and a feature map $G$ makes $\rm {P}_S \left( \mathcal {M}(\rm {X}))=\rm {P}_T(\mathcal {M}(\rm {X}) \right)$, then $\rm {P}_S(\rm {Y}=i|\mathcal {M}(\rm {X}))=\rm {P}_S(\rm {Y}=i)$. Proofs appear in Appendix A. ### Problem of Domain-Invariant Representation Learning ::: Remark.
According to Theorem UNKREF8, we know that when $\rm {P}_S(\rm {X}|\rm {Y})=\rm {P}_T(\rm {X}|\rm {Y})$ and $\rm {P}_S(\rm {Y}=i) \ne \rm {P}_T(\rm {Y}=i)$, forcing $G(\rm {X})$ to be domain-invariant inclines to make data of class $i$ mix with data of other classes in the space of $G(\rm {X})$. This will make it difficult for the supervised classifier to distinguish inputs of class $i$ from inputs of the other classes. Think about such an extreme case that every instance $x$ is mapped to a consistent point $g_0$ in $G(\rm {X})$. In this case, $\rm {P}_S(G(\rm {X})=g_0)= \rm {P}_T(G(\rm {X})=g_0) = 1$. Therefore, $G(\rm {X})$ is domain-invariant. As a result, the supervised classifier will assign the label $y^* = \operatornamewithlimits{arg\,max}_y \rm {P}_S(\rm {Y}=y)$ to all input examples. This is definitely unacceptable. To give a more intuitive illustration of the above analysis, we offer several empirical studies on Theorem UNKREF8 in Appendix B. When $\rm {P}_S(\rm {Y})\ne \rm {P}_T(\rm {Y})$ and $\rm {P}_S(\rm {X}|\rm {Y}) \ne \rm {P}_T(\rm {X}|\rm {Y})$, we did not obtain such a strong conclusion as Theorem UNKREF8. Instead, we deduced a conflict between the object of achieving superior classification performance and that of making features domain-invariant. Suppose that $\rm {P}_S(\rm {Y}=i) \ne \rm {P}_T(\rm {Y}=i)$ and instances of class $i$ are completely distinguishable from instances of the rest classes in $G(\rm {X})$, i.e.,: In DIRL, we hope that: Consider the region $x \in \mathcal {X}_i$, where $\rm {P}(G(\rm {X}=x)|\rm {Y}=i)>0$. According to the above assumption, we know that $\rm {P}(G(\rm {X}=x \in \mathcal {X}_i)|\rm {Y} \ne i) = 0$. Therefore, applying DIRL will force in region $x \in \mathcal {X}_i$. Taking the integral of $x$ over $\mathcal {X}_i$ for both sides of the equation, we have $\rm {P}_S(\rm {Y}=i) = \rm {P}_T(\rm {Y}=i)$. This deduction contradicts with the setting that $\rm {P}_S(\rm {Y}=i) \ne \rm {P}_T(\rm {Y}=i)$. Therefore, $G(\rm {X})$ is impossible fully class-separable when it is domain-invariant. Note that the object of the supervised learning is exactly to make $G(\rm {X})$ class-separable. Thus, this actually indicates a conflict between the supervised learning and the domain-invariant representation learning. Based on the above analysis, we can conclude that it is impossible to obtain a feature representation $G(X)$ that is class-separable and at the same time, domain-invariant using the DIRL framework, when $\rm {P}(\rm {Y})$ shifts across domains. However, the shift of $\rm {P}(\rm {Y})$ can exist in many cross-domain sentiment analysis tasks. Therefore, it is worthy of studying in order to deal with the problem of DIRL. ### Weighted Domain Invariant Representation Learning
According to the above analysis, we proposed a weighted version of DIRL to address the problem caused by the shift of $\rm {P}(\rm {Y})$ to DIRL. The key idea of this framework is to first align $\rm {P}(\rm {Y})$ across domains before performing domain-invariant learning, and then take account the shift of $\rm {P}(\rm {Y})$ in the label prediction procedure. Specifically, it introduces a class weight $\mathbf {w}$ to weigh source domain examples by class. Based on the weighted source domain, the domain shift problem is resolved in two steps. In the first step, it applies DIRL on the target domain and the weighted source domain, aiming to alleviate the influence of the shift of $\rm {P}(\rm {Y})$ during the alignment of $\rm {P}(\rm {X}|\rm {Y})$. In the second step, it uses $\mathbf {w}$ to reweigh the supervised classifier $\rm {P}_S(\rm {Y}|\rm {X})$ obtained in the first step for target domain label prediction. We detail these two steps in §SECREF10 and §SECREF14, respectively. ### Weighted Domain Invariant Representation Learning ::: Align @!START@$\rm {P}(\rm {X}|\rm {Y})$@!END@ with Class Weight
The motivation behind this practice is to adjust data distribution of the source domain or the target domain to alleviate the shift of $\rm {P}(\rm {Y})$ across domains before applying DIRL. Consider that we only have labels of source domain data, we choose to adjust data distribution of the source domain. To achieve this purpose, we introduce a trainable class weight $\mathbf {w}$ to reweigh source domain examples by class when performing DIRL, with $\mathbf {w}_i > 0$. Specifically, we hope that: and we denote $\mathbf {w}^*$ the value of $\mathbf {w}$ that makes this equation hold. We shall see that when $\mathbf {w}=\mathbf {w}^*$, DIRL is to align $\rm {P}_S(G(\rm {X})|\rm {Y})$ with $\rm {P}_T(G(\rm {X})|\rm {Y})$ without the shift of $\rm {P}(\rm {Y})$. According to our analysis, we know that due to the shift of $\rm {P}(\rm {Y})$, there is a conflict between the training objects of the supervised learning $\mathcal {L}_{sup}$ and the domain-invariant learning $\mathcal {L}_{inv}$. And the conflict degree will decrease as $\rm {P}_S(\rm {Y})$ getting close to $\rm {P}_T(\rm {Y})$. Therefore, during model training, $\mathbf {w}$ is expected to be optimized toward $\mathbf {w}^*$ since it will make $\rm {P}(\rm {Y})$ of the weighted source domain close to $\rm {P}_T(\rm {Y})$, so as to solve the conflict. We now show how to transfer existing DIRL models to their WDIRL counterparts with the above idea. Let $\mathbb {S}:\rm {P} \rightarrow {R}$ denote a statistic function defined over a distribution $\rm {P}$. For example, the expectation function $\mathbb {E}(\rm {X})$ in $\mathbb {E}(\rm {X}_S) \equiv \mathbb {E}(\rm {X})(\rm {P}_S(\rm {X}))$ is a concrete instaintiation of $\mathbb {S}$. In general, to transfer models from DIRL to WDIRL, we should replace $\mathbb {S}(\rm {P}_S(\rm {X}))$ defined in $\mathcal {L}_{inv}$ with Take the CMD metric as an example. In WDIRL, the revised form of ${\text{CMD}}_K$ is defined by: Here, $\mathbb {E}(\rm {X}_S|\rm {Y}_S=i) \equiv \mathbb {E}(\rm {X})(\rm {P}_S(\rm {X}|\rm {Y}=i))$ denotes the expectation of $\rm {X}$ over distribution $\rm {P}_S(\rm {X}|\rm {Y}=i)$. Note that both $\rm {P}_S(\rm {Y}=i)$ and $\mathbb {E}(\rm {X}_S|\rm {Y}_S=i)$ can be estimated using source labeled data, and $\mathbb {E}(\rm {X}_T)$ can be estimated using target unlabeled data. As for those adversarial-learning-based DIRL methods, e.g., DANN BIBREF2, the revised domain-invariant loss can be precisely defined by: During model training, $D$ is optimized in the direction to minimize $\hat{\mathcal {L}}_d$, while $G$ and $\mathbf {w}$ are optimized to maximize $\hat{\mathcal {L}}_d$. In the following, we denote $\widehat{\text{JSD}}(\rm {P}_S, \rm {P}_T)$ the equivalent loss defined over $G$ for the revised version of domain adversarial learning. The general task loss in WDIRL is defined by: where $\hat{\mathcal {L}}_{inv}$ is a unified representation of the domain-invariant loss in WDIRL, such as $\widehat{\text{CMD}}_K$ and $\widehat{\text{JSD}}(\rm {P}_S, \rm {P}_T)$. ### Weighted Domain Invariant Representation Learning ::: Align @!START@$\rm {P}(\rm {Y}|\rm {X})$@!END@ with Class Weight
In the above step, we align $\rm {P}(\rm {X}|\rm {Y})$ across domains by performing domain-invariant learning on the class-weighted source domain and the original target domain. In this step, we deal with the shift of $\rm {P}(\rm {Y})$. Suppose that we have successfully resolved the shift of $\rm {P}(\rm {X}|\rm {Y})$ with $G$, i.e., $\rm {P}_S(G(\rm {X})|\rm {Y})=\rm {P}_T(G(\rm {X})|\rm {Y})$. Then, according to the work of BIBREF29, we have: where $\gamma (\rm {Y}=i)={\rm {P}_T(\rm {Y}=i)}/{\rm {P}_S(\rm {Y}=i)}$. Of course, in most of the real-world tasks, we do not know the value of $\gamma (\rm {Y}=i)$. However, note that $\gamma (\rm {Y}=i)$ is exactly the expected class weight $\mathbf {w}^*_i$. Therefore, a natural practice of this step is to estimate $\gamma (\rm {Y}=i)$ with the obtained $\mathbf {w}_i$ in the first step and estimate $\rm {P}_T(\rm {Y}|G(\rm {X}))$ with: In summary, to transfer methods of the DIRL paradigm to WDIRL, we should: first revise the definition of $\mathcal {L}_{inv}$, obtaining its corresponding WDIRL form $\hat{\mathcal {L}}_{inv}$; then perform supervised learning and domain-invariant representation learning on $\mathcal {D}_S$ and $\mathcal {D}_T$ according to Eq. (DISPLAY_FORM13), obtaining a supervised classifier $\rm {P}_S(\rm {Y}|\rm {X}; \mathbf {\Phi })$ and a class weight vector $\mathbf {w}$; and finally, adjust $\rm {P}_S(\rm {Y}|\rm {X}; \mathbf {\Phi })$ using $\mathbf {w}$ according to Eq. (DISPLAY_FORM16) and obtain the target domain classifier $\rm {P}_T(\rm {Y}|\rm {X}; \mathbf {\Phi })$. ### Experiment ::: Experiment Design
Through the experiments, we empirically studied our analysis on DIRL and the effectiveness of our proposed solution in dealing with the problem it suffered from. In addition, we studied the impact of each step described in §SECREF10 and §SECREF14 to our proposed solution, respectively. To performe the study, we carried out performance comparison between the following models: SO: the source-only model trained using source domain labeled data without any domain adaptation. CMD: the centre-momentum-based domain adaptation model BIBREF3 of the original DIRL framework that implements $\mathcal {L}_{inv}$ with $\text{CMD}_K$. DANN: the adversarial-learning-based domain adaptation model BIBREF2 of the original DIRL framework that implements $\mathcal {L}_{inv}$ with $\text{JSD}(\rm {P}_S, \rm {P}_T)$. $\text{CMD}^\dagger $: the weighted version of the CMD model that only applies the first step (described in §SECREF10) of our proposed method. $\text{DANN}^\dagger $: the weighted version of the DANN model that only applies the first step of our proposed method. $\text{CMD}^{\dagger \dagger }$: the weighted version of the CMD model that applies both the first and second (described in §SECREF14) steps of our proposed method. $\text{DANN}^{\dagger \dagger }$: the weighted version of the DANN model that applies both the first and second steps of our proposed method. $\text{CMD}^{*}$: a variant of $\text{CMD}^{\dagger \dagger }$ that assigns $\mathbf {w}^*$ (estimate from target labeled data) to $\mathbf {w}$ and fixes this value during model training. $\text{DANN}^{*}$: a variant of $\text{DANN}^{\dagger \dagger }$ that assigns $\mathbf {w}^*$ to $\mathbf {w}$ and fixes this value during model training. Intrinsically, SO can provide an empirical lowerbound for those domain adaptation methods. $\text{CMD}^{*}$ and $\text{DANN}^{*}$ can provide the empirical upbound of $\text{CMD}^{\dagger \dagger }$ and $\text{DANN}^{\dagger \dagger }$, respectively. In addition, by comparing performance of $\text{CMD}^{*}$ and $\text{DANN}^{*}$ with that of $\text{SO}$, we can know the effectiveness of the DIRL framework when $\rm {P}(\rm {Y})$ dose not shift across domains. By comparing $\text{CMD}^\dagger $ with $\text{CMD}$, or comparing $\text{DANN}^\dagger $ with $\text{DANN}$, we can know the effectiveness of the first step of our proposed method. By comparing $\text{CMD}^{\dagger \dagger }$ with $\text{CMD}^{\dagger }$, or comparing $\text{DANN}^{\dagger \dagger }$ with $\text{DANN}^{\dagger }$, we can know the impact of the second step of our proposed method. And finally, by comparing $\text{CMD}^{\dagger \dagger }$ with $\text{CMD}$, or comparing $\text{DANN}^{\dagger \dagger }$ with $\text{DANN}$, we can know the general effectiveness of our proposed solution. ### Experiment ::: Dataset and Task Design
We conducted experiments on the Amazon reviews dataset BIBREF9, which is a benchmark dataset in the cross-domain sentiment analysis field. This dataset contains Amazon product reviews of four different product domains: Books (B), DVD (D), Electronics (E), and Kitchen (K) appliances. Each review is originally associated with a rating of 1-5 stars and is encoded in 5,000 dimensional feature vectors of bag-of-words unigrams and bigrams. ### Experiment ::: Dataset and Task Design ::: Binary-Class.
From this dataset, we constructed 12 binary-class cross-domain sentiment analysis tasks: B$\rightarrow $D, B$\rightarrow $E, B$\rightarrow $K, D$\rightarrow $B, D$\rightarrow $E, D$\rightarrow $K, E$\rightarrow $B, E$\rightarrow $D, E$\rightarrow $K, K$\rightarrow $B, K$\rightarrow $D, K$\rightarrow $E. Following the setting of previous works, we treated a reviews as class `1' if it was ranked up to 3 stars, and as class `2' if it was ranked 4 or 5 stars. For each task, $\mathcal {D}_S$ consisted of 1,000 examples of each class, and $\mathcal {D}_T$ consists of 1500 examples of class `1' and 500 examples of class `2'. In addition, since it is reasonable to assume that $\mathcal {D}_T$ can reveal the distribution of target domain data, we controlled the target domain testing dataset to have the same class ratio as $\mathcal {D}_T$. Using the same label assigning mechanism, we also studied model performance over different degrees of $\rm {P}(\rm {Y})$ shift, which was evaluated by the max value of $\rm {P}_S(\rm {Y}=i)/\rm {P}_T(\rm {Y}=i), \forall i=1, \cdots , L$. Please refer to Appendix C for more detail about the task design for this study. ### Experiment ::: Dataset and Task Design ::: Multi-Class.
We additionally constructed 12 multi-class cross-domain sentiment classification tasks. Tasks were designed to distinguish reviews of 1 or 2 stars (class 1) from those of 4 stars (class 2) and those of 5 stars (class 3). For each task, $\mathcal {D}_S$ contained 1000 examples of each class, and $\mathcal {D}_T$ consisted of 500 examples of class 1, 1500 examples of class 2, and 1000 examples of class 3. Similarly, we also controlled the target domain testing dataset to have the same class ratio as $\mathcal {D}_T$. ### Experiment ::: Implementation Detail
For all studied models, we implemented $G$ and $f$ using the same architectures as those in BIBREF3. For those DANN-based methods (i.e., DANN, $\text{DANN}^{\dagger }$, $\text{DANN}^{\dagger \dagger }$, and $\text{DANN}^{*}$), we implemented the discriminator $D$ using a 50 dimensional hidden layer with relu activation functions and a linear classification layer. Hyper-parameter $K$ of $\text{CMD}_K$ and $\widehat{\text{CMD}}_K$ was set to 5 as suggested by BIBREF3. Model optimization was performed using RmsProp BIBREF30. Initial learning rate of $\mathbf {w}$ was set to 0.01, while that of other parameters was set to 0.005 for all tasks. Hyper-parameter $\alpha $ was set to 1 for all of the tested models. We searched for this value in range $\alpha =[1, \cdots , 10]$ on task B $\rightarrow $ K. Within the search, label distribution was set to be uniform, i.e., $\rm {P}(\rm {Y}=i)=1/L$, for both domain B and K. We chose the value that maximize the performance of CMD on testing data of domain K. You may notice that this practice conflicts with the setting of unsupervised domain adaptation that we do not have labeled data of the target domain for training or developing. However, we argue that this practice would not make it unfair for model comparison since all of the tested models shared the same value of $\alpha $ and $\alpha $ was not directly fine-tuned on any tested task. With the same consideration, for every tested model, we reported its best performance achieved on testing data of the target domain during its training. To initialize $\mathbf {w}$, we used label prediction of the source-only model. Specifically, let $\rm {P}_{SO}(\rm {Y}|\rm {X}; \mathbf {\theta }_{SO})$ denote the trained source-only model. We initialized $\mathbf {w}_i$ by: Here, $\mathbb {I}$ denotes the indication function. To offer an intuitive understanding to this strategy, we report performance of WCMD$^{\dagger \dagger }$ over different initializations of $\mathbf {w}$ on 2 within-group (B$\rightarrow $D, E$\rightarrow $K) and 2 cross-group (B$\rightarrow $K, D$\rightarrow $E) binary-class domain adaptation tasks in Figure FIGREF33. Here, we say that domain B and D are of a group, and domain E and K are of another group since B and D are similar, as are E and K, but the two groups are different from one another BIBREF9. Note that $\rm {P}_{S}(\rm {Y}=1)=0.5$ is a constant, which is estimated using source labeled data. From the figure, we can obtain three main observations. First, WCMD$^{\dagger \dagger }$ generally outperformed its CMD counterparts with different initialization of $\mathbf {w}$. Second, it was better to initialize $\mathbf {w}$ with a relatively balanced value, i.e., $\mathbf {w}_i \rm {P}_S(\rm {Y}=i) \rightarrow \frac{1}{L}$ (in this experiment, $L=2$). Finally, $\mathbf {w}^0$ was often a good initialization of $\mathbf {w}$, indicating the effectiveness of the above strategy. ### Experiment ::: Main Result
Table TABREF27 shows model performance on the 12 binary-class cross-domain tasks. From this table, we can obtain the following observations. First, CMD and DANN underperform the source-only model (SO) on all of the 12 tested tasks, indicating that DIRL in the studied situation will degrade the domain adaptation performance rather than improve it. This observation confirms our analysis. Second, $\text{CMD}^{\dagger \dagger }$ consistently outperformed CMD and SO. This observation shows the effectiveness of our proposed method for addressing the problem of the DIRL framework in the studied situation. Similar conclusion can also be obtained by comparing performance of $\text{DANN}^{\dagger \dagger }$ with that of DANN and SO. Third, $\text{CMD}^{\dagger }$ and $\text{DANN}^{\dagger }$ consistently outperformed $\text{CMD}$ and DANN, respectively, which shows the effectiveness of the first step of our proposed method. Finally, on most of the tested tasks, $\text{CMD}^{\dagger \dagger }$ and $\text{DANN}^{\dagger \dagger }$ outperforms $\text{CMD}^{\dagger }$ and $\text{DANN}^{\dagger }$, respectively. Figure FIGREF35 depicts the relative improvement, e.g., $(\text{Acc}(\text{CMD})-\text{Acc}(\text{SO}))/\text{Acc}(\text{SO})$, of the domain adaptation methods over the SO baseline under different degrees of $\rm {P}(\rm {Y})$ shift, on two binary-class domain adaptation tasks (You can refer to Appendix C for results of the other models on other tasks). From the figure, we can see that the performance of CMD generally got worse as the increase of $\rm {P}(\rm {Y})$ shift. In contrast, our proposed model $\text{CMD}^{\dagger \dagger }$ performed robustly to the varying of $\rm {P}(\rm {Y})$ shift degree. Moreover, it can achieve the near upbound performance characterized by $\text{CMD}^{*}$. This again verified the effectiveness of our solution. Table TABREF34 reports model performance on the 2 within-group (B$\rightarrow $D, E$\rightarrow $K) and the 2 cross-group (B$\rightarrow $K, D$\rightarrow $E) multi-class domain adaptation tasks (You can refer to Appendix D for results on the other tasks). From this table, we observe that on some tested tasks, $\text{CMD}^{\dagger \dagger }$ and $\text{DANN}^{\dagger \dagger }$ did not greatly outperform or even slightly underperformed $\text{CMD}^{\dagger }$ and $\text{DANN}^{\dagger }$, respectively. A possible explanation of this phenomenon is that the distribution of $\mathcal {D}_T$ also differs from that of the target domain testing dataset. Therefore, the estimated or learned value of $\mathbf {w}$ using $\mathcal {D}_T$ is not fully suitable for application to the testing dataset. This explanation is verified by the observation that $\text{CMD}^{\dagger }$ and $\text{DANN}^{\dagger }$ also slightly outperforms $\text{CMD}^{*}$ and $\text{DANN}^{*}$ on these tasks, respectively. ### Conclusion
In this paper, we studied the problem of the popular domain-invariant representation learning (DIRL) framework for domain adaptation, when $\rm {P}(\rm {Y})$ changes across domains. To address the problem, we proposed a weighted version of DIRL (WDIRL). We showed that existing methods of the DIRL framework can be easily transferred to our WDIRL framework. Extensive experimental studies on benchmark cross-domain sentiment analysis datasets verified our analysis and showed the effectiveness of our proposed solution. Table 1: Mean accuracy ± standard deviation over five runs on the 12 binary-class cross-domain tasks. Figure 1: Mean accuracy of WCMD†† over different initialization of w. The empirical optimum value of w makes w1PS(Y = 1) = 0.75. The dot line in the same color denotes performance of the CMD model and ‘w0’ annotates performance of WCMD†† when initializing w with w0. Figure 2: Relative improvement over the SO baseline under different degrees of P(Y) shift on the B→D and B →K binary-class domain adaptation tasks. Table 2: Mean accuracy ± standard deviation over five runs on the 2 within-group and 2 cross-group multiclass domain-adaptation tasks. | Through the experiments, we empirically studied our analysis on DIRL and the effectiveness of our proposed solution in dealing with the problem it suffered from. |
Had the truck driver driving along Route 202 not noticed the change in road ahead while traveling, what would have likely happened?
A. He would have driven down into the pit where Superior was formerly located.
B. He would have passed right over the town and missed it totally.
C. He would have spilled his coffee while trying to make the sudden stop.
D. He would have floated above the ground and continued driving into the town of Superior.
| And Then the Town Took Off by RICHARD WILSON ACE BOOKS, INC. 23 West 47th Street, New York 36, N.Y. AND THEN THE TOWN TOOK OFF Copyright ©, 1960, by Ace Books, Inc. All Rights Reserved For Felicitas K. Wilson THE SIOUX SPACEMAN Copyright ©, 1960, by Ace Books, Inc. Printed in U.S.A. THE CITY THAT RAN OFF THE MAP The town of Superior, Ohio, certainly was living up to its name! In what was undoubtedly the most spectacular feat of the century, it simply picked itself up one night and rose two full miles above Earth! Radio messages stated simply that Superior had seceded from Earth. But Don Cort, stranded on that rising town, was beginning to suspect that nothing was simple about Superior except its citizens. Calmly they accepted their rise in the world as being due to one of their local townspeople, a crackpot professor. But after a couple of weeks of floating around, it began to be obvious that the professor had no idea how to get them down. So then it was up to Cort: either find a way to anchor Superior, or spend the rest of his days on the smallest—and the nuttiest—planet in the galaxy! I The town of Superior, Ohio, disappeared on the night of October 31. A truck driver named Pierce Knaubloch was the first to report it. He had been highballing west along Route 202, making up for the time he'd spent over a second cup of coffee in a diner, when he screeched to a stop. If he'd gone another twenty-five feet he'd have gone into the pit where Superior had been. Knaubloch couldn't see the extent of the pit because it was too dark, but it looked big. Bigger than if a nitro truck had blown up, which was his first thought. He backed up two hundred feet, set out flares, then sped off to a telephone. The state police converged on the former site of Superior from several directions. Communicating by radiophone across the vast pit, they confirmed that the town undoubtedly was missing. They put in a call to the National Guard. The guard surrounded the area with troops—more than a thousand were needed—to keep people from falling into the pit. A pilot who flew over it reported that it looked as if a great ice-cream scoop had bitten into the Ohio countryside. The Pennsylvania Railroad complained that one of its passenger trains was missing. The train's schedule called for it to pass through but not stop at Superior at 11:58. That seemed to fix the time of the disappearance at midnight. The truck driver had made his discovery shortly after midnight. Someone pointed out that October 31 was Halloween and that midnight was the witching hour. Somebody else said nonsense, they'd better check for radiation. A civil defense official brought up a Geiger counter, but no matter how he shook it and rapped on it, it refused to click. A National Guard officer volunteered to take a jeep down into the pit, having found a spot that seemed navigable. He was gone a long time but when he came out the other side he reported that the pit was concave, relatively smooth, and did not smell of high explosives. He'd found no people, no houses—no sign of anything except the pit itself. The Governor of Ohio asked Washington whether any unidentified planes had been over the state. Washington said no. The Pentagon and the Atomic Energy Commission denied that they had been conducting secret experiments. Nor had there been any defense plants in Superior that might have blown up. The town's biggest factory made kitchen sinks and the next biggest made bubble gum. A United Airlines pilot found Superior early on the morning of November 1. The pilot, Captain Eric Studley, who had never seen a flying saucer and hoped never to see one, was afraid now that he had. The object loomed out of a cloudbank at twelve thousand feet and Studley changed course to avoid it. He noted with only minimum satisfaction that his co-pilot also saw the thing and wondered why it wasn't moving at the terrific speed flying saucers were allegedly capable of. Then he saw the church steeple on it. A few minutes later he had relayed a message from Superior, formerly of Ohio, addressed to whom it might concern: It said that Superior had seceded from Earth. One other radio message came from Superior, now airborne, on that first day. A ham radio operator reported an unidentified voice as saying plaintively: " Cold up here!" Don Cort had been dozing in what passed for the club car on the Buckeye Cannonball when the train braked to a stop. He looked out the window, hoping this was Columbus, where he planned to catch a plane east. But it wasn't Columbus. All he could see were some lanterns jogging as trainmen hurried along the tracks. The conductor looked into the car. The redhead across the aisle in whom Don had taken a passing interest earlier in the evening asked, "Why did we stop?" "Somebody flagged us down," the conductor said. "We don't make a station stop at Superior on this run." The girl's hair was a subtle red, but false. When Don had entered the club car he'd seen her hatless head from above and noticed that the hair along the part was dark. Her eyes had been on a book and Don had the opportunity for a brief study of her face. The cheeks were full and untouched by make-up. There were lines at the corners of her mouth which indicated a tendency to arrange her expression into one of disapproval. The lips were full, like the cheeks, but it was obvious that the scarlet lipstick had contrived a mouth a trifle bigger than the one nature had given her. Her glance upward at that moment interrupted his examination, which had been about to go on to her figure. Later, though, he was able to observe that it was more than adequate. If the girl had given Don Cort more than that one glance, or if it had been a trained, all-encompassing glance, she would have seen a man in his mid-twenties—about her age—lean, tall and straight-shouldered, with once-blond hair now verging on dark brown, a face neither handsome nor ugly, and a habit of drawing the inside of his left cheek between his teeth and nibbling at it thoughtfully. But it was likely that all she noticed then was the brief case he carried, attached by a chain to a handcuff on his left wrist. "Will we be here long?" Don asked the conductor. He didn't want to miss his plane at Columbus. The sooner he got to Washington, the sooner he'd get rid of the brief case. The handcuff it was attached to was one reason why his interest in the redhead had been only passing. "Can't say," the conductor told him. He let the door close again and went down to the tracks. Don hesitated, shrugged at the redhead, said, "Excuse me," and followed the conductor. About a dozen people were milling around the train as it sat in the dark, hissing steam. Don made his way up to the locomotive and found a bigger knot of people gathered in front of the cowcatcher. Some sort of barricade had been put up across the tracks and it was covered with every imaginable kind of warning device. There were red lanterns, both battery and electric; flashlights; road flares; and even an old red shirt. Don saw two men who must have been the engineer and the fireman talking to an old bearded gentleman wearing a civil defense helmet, a topcoat and riding boots. "You'd go over the edge, I tell you," the old gentleman was saying. "If you don't get this junk off the line," the engineer said, "I'll plow right through it. Off the edge! you crazy or something?" "Look for yourself," the old man in the white helmet said. "Go ahead. Look." The engineer was exasperated. He turned to the fireman. "You look. Humor the old man. Then let's go." The bearded man—he called himself Professor Garet—went off with the fireman. Don followed them. They had tramped a quarter of a mile along the gravel when the fireman stopped. "Okay," he said "where's the edge? I don't see nothing." The tracks seemed to stretch forever into the darkness. "It's another half mile or so," the professor said. "Well, let's hurry up. We haven't got all night." The old man chuckled. "I'm afraid you have." They came to it at last, stopping well back from it. Professor Garet swelled with pride, it seemed, as he made a theatrical gesture. "Behold," he said. "Something even Columbus couldn't find. The edge of the world." True, everything seemed to stop, and they could see stars shining low on the horizon where stars could not properly be expected to be seen. Don Cort and the fireman walked cautiously toward the edge while the professor ambled ahead with the familiarity of one who had been there before. But there was a wind and they did not venture too close. Nevertheless, Don could see that it apparently was a neat, sharp edge, not one of your old ragged, random edges such as might have been caused by an explosion. This one had the feeling of design behind it. Standing on tiptoe and repressing a touch of giddiness, Don looked over the edge. He didn't have to stand on tiptoe any more than he had to sit on the edge of his seat during the exciting part of a movie, but the situation seemed to call for it. Over the edge could be seen a big section of Ohio. At least he supposed it was Ohio. Don looked at the fireman, who had an unbelieving expression on his face, then at the bearded old man, who was smiling and nodding. "You see what I mean," he said. "You would have gone right over. I believe you would have had a two-mile fall." "Of course you could have stayed aboard the train," the man driving the old Pontiac said, "but I really think you'll be more comfortable at Cavalier." Don Cort, sitting in the back seat of the car with the redhead from the club car, asked, "Cavalier?" "The college. The institute, really; it's not accredited. What did you say your name was, miss?" "Jen Jervis," she said. "Geneva Jervis, formally." "Miss Jervis. I'm Civek. You know Mr. Cort, I suppose." The girl smiled sideways. "We have a nodding acquaintance." Don nodded and grinned. "There's plenty of room in the dormitories," Civek said. "People don't exactly pound on the gates and scream to be admitted to Cavalier." "Are you connected with the college?" Don asked. "Me? No. I'm the mayor of Superior. The old town's really come up in the world, hasn't it?" "Overnight," Geneva Jervis said. "If what Mr. Cort and the fireman say is true. I haven't seen the edge myself." "You'll have a better chance to look at it in the morning," the mayor said, "if we don't settle back in the meantime." "Was there any sort of explosion?" Don asked. "No. There wasn't any sensation at all, as far as I noticed. I was watching the late show—or trying to. My house is down in a hollow and reception isn't very good, especially with old English movies. Well, all of a sudden the picture sharpened up and I could see just as plain. Then the phone rang and it was Professor Garet." "The old fellow with the whiskers and the riding boots?" Jen Jervis asked. "Yes. Osbert Garet, Professor of Magnology at the Cavalier Institute of Applied Sciences." "Professor of what?" "Magnology. As I say, the school isn't accredited. Well, Professor Garet telephoned and said, 'Hector'—that's my name, Hector Civek—'everything's up in the air.' He was having his little joke, of course. I said, 'What?' and then he told me." "Told you what?" Jen Jervis asked. "I mean, does he have any theory about it?" "He has a theory about everything. I think what he was trying to convey was that this—this levitation confirmed his magnology principle." "What's that?" Don asked. "I haven't the faintest idea. I'm a politician, not a scientist. Professor Garet went on about it for a while, on the telephone, about magnetism and gravity, but I think he was only calling as a courtesy, so the mayor wouldn't look foolish the next morning, not knowing his town had flown the coop." "What's the population of Superior?" "Three thousand, including the students at the institute. Three thousand and forty, counting you people from the train. I guess you'll be with us for a while." "What do you mean by that?" Jen Jervis asked. "Well, I don't see how you can get down. Do you?" "Does Superior have an airport?" Don asked. "I've got to get back to—to Earth." It sounded odd to put it that way. "Nope," Civek said. "No airport. No place for a plane to land, either." "Maybe not a plane," Don said, "but a helicopter could land just about anywhere." "No helicopters here, either." "Maybe not. But I'll bet they're swarming all over you by morning." "Hm," said Hector Civek. Don couldn't quite catch his expression in the rearview mirror. "I suppose they could, at that. Well, here's Cavalier. You go right in that door, where the others are going. There's Professor Garet. I've got to see him—excuse me." The mayor was off across the campus. Don looked at Geneva Jervis, who was frowning. "Are you thinking," he asked, "that Mayor Civek was perhaps just a little less than completely honest with us?" "I'm thinking," she said, "that I should have stayed with Aunt Hattie another night, then taken a plane to Washington." "Washington?" Don said. "That's where I'm going. I mean where I was going before Superior became airborne. What do you do in Washington, Miss Jervis?" "I work for the Government. Doesn't everybody?" "Not everybody. Me, for instance." "No?" she said. "Judging by that satchel you're handcuffed to, I'd have thought you were a courier for the Pentagon. Or maybe State." He laughed quickly and loudly because she was getting uncomfortably close. "Oh, no. Nothing so glamorous. I'm a messenger for the Riggs National Bank, that's all. Where do you work?" "I'm with Senator Bobby Thebold, S.O.B." Don laughed again. "He sure is." " Mister Cort!" she said, annoyed. "You know as well as I do that S.O.B. stands for Senate Office Building. I'm his secretary." "I'm sorry. We'd better get out and find a place to sleep. It's getting late." " Places to sleep," she corrected. She looked angry. "Of course," Don said, puzzled by her emphasis. "Come on. Where they put you, you'll probably be surrounded by co-eds, even if I could get out of this cuff." He took her bag in his free hand and they were met by a gray-haired woman who introduced herself as Mrs. Garet. "We'll try to make you comfortable," she said. "What a night, eh? The professor is simply beside himself. We haven't had so much excitement since the cosmolineator blew up." They had a glimpse of the professor, still in his CD helmet, going around a corner, gesticulating wildly to someone wearing a white laboratory smock. II Don Cort had slept, but not well. He had tried to fold the brief case to pull it through his sleeve so he could take his coat off, but whatever was inside the brief case was too big. Cavalier had given him a room to himself at one end of a dormitory and he'd taken his pants off but had had to sleep with his coat and shirt on. He got up, feeling gritty, and did what little dressing was necessary. It was eight o'clock, according to the watch on the unhandcuffed wrist, and things were going on. He had a view of the campus from his window. A bright sun shone on young people moving generally toward a squat building, and other people going in random directions. The first were students going to breakfast, he supposed, and the others were faculty members. The air was very clear and the long morning shadows distinct. Only then did he remember completely that he and the whole town of Superior were up in the air. He went through the dormitory. A few students were still sleeping. The others had gone from their unmade beds. He shivered as he stepped outdoors. It was crisp, if not freezing, and his breath came out visibly. First he'd eat, he decided, so he'd be strong enough to go take a good look over the edge, in broad daylight, to the Earth below. The mess hall, or whatever they called it, was cafeteria style and he got in line with a tray for juice, eggs and coffee. He saw no one he knew, but as he was looking for a table a willowy blonde girl smiled and gestured to the empty place opposite her. "You're Mr. Cort," she said. "Won't you join me?" "Thanks," he said, unloading his tray. "How did you know?" "The mystery man with the handcuff. You'd be hard to miss. I'm Alis—that's A-l-i-s, not A-l-i-c-e—Garet. Are you with the FBI? Or did you escape from jail?" "How do you do. No, just a bank messenger. What an unusual name. Professor Garet's daughter?" "The same," she said. "Also the only. A pity, because if there'd been two of us I'd have had a fifty-fifty chance of going to OSU. As it is, I'm duty-bound to represent the second generation at the nut factory." "Nut factory? You mean Cavalier?" Don struggled to manipulate knife and fork without knocking things off the table with his clinging brief case. "Here, let me cut your eggs for you," Alis said. "You'd better order them scrambled tomorrow. Yes, Cavalier. Home of the crackpot theory and the latter-day alchemist." "I'm sure it's not that bad. Thanks. As for tomorrow, I hope to be out of here by then." "How do you get down from an elephant? Old riddle. You don't; you get down from ducks. How do you plan to get down from Superior?" "I'll find a way. I'm more interested at the moment in how I got up here." "You were levitated, like everybody else." "You make it sound deliberate, Miss Garet, as if somebody hoisted a whole patch of real estate for some fell purpose." "Scarcely fell , Mr. Cort. As for it being deliberate, that seems to be a matter of opinion. Apparently you haven't seen the papers." "I didn't know there were any." "Actually there's only one, the Superior Sentry , a weekly. This is an extra. Ed Clark must have been up all night getting it out." She opened her purse and unfolded a four-page tabloid. Don blinked at the headline: Town Gets High "Ed Clark's something of an eccentric, like everybody else in Superior," Alis said. Don read the story, which seemed to him a capricious treatment of an apparently grave situation. Residents having business beyond the outskirts of town today are advised not to. It's a long way down. Where Superior was surrounded by Ohio, as usual, today Superior ends literally at the town line. A Citizens' Emergency Fence-Building Committee is being formed, but in the meantime all are warned to stay well away from the edge. The law of gravity seems to have been repealed for the town but it is doubtful if the same exemption would apply to a dubious individual bent on investigating.... Don skimmed the rest. "I don't see anything about it being deliberate." Alis had been creaming and sugaring Don's coffee. She pushed it across to him and said, "It's not on page one. Ed Clark and Mayor Civek don't get along, so you'll find the mayor's statement in a box on page three, bottom." Don creased the paper the other way, took a sip of coffee, nodded his thanks, and read: Mayor Claims Secession From Earth Mayor Hector Civek, in a proclamation issued locally by hand and dropped to the rest of the world in a plastic shatter-proof bottle, said today that Superior has seceded from Earth. His reasons were as vague as his explanation. The "reasons" include these: (1) Superior has been discriminated against by county, state and federal agencies; (2) Cavalier Institute has been held up to global derision by orthodox (presumably meaning accredited) colleges and universities; and (3) chicle exporters have conspired against the Superior Bubble Gum Company by unreasonably raising prices. The "explanation" consists of a 63-page treatise on applied magnology by Professor Osbert Garet of Cavalier which the editor (a) does not understand; (b) lacks space to publish; and which (it being atrociously handwritten) he (c) has not the temerity to ask his linotype operator to set. Don said, "I'm beginning to like this Ed Clark." "He's a doll," Alis said. "He's about the only one in town who stands up to Father." "Does your father claim that he levitated Superior off the face of the Earth?" "Not to me he doesn't. I'm one of those banes of his existence, a skeptic. He gave up trying to magnolize me when I was sixteen. I had a science teacher in high school—not in Superior, incidentally—who gave me all kinds of embarrassing questions to ask Father. I asked them, being a natural-born needler, and Father has disowned me intellectually ever since." "How old are you, Miss Garet, if I may ask?" She sat up straight and tucked her sweater tightly into her skirt, emphasizing her good figure. To a male friend Don would have described the figure as outstanding. She had mocking eyes, a pert nose and a mouth of such moist red softness that it seemed perpetually waiting to be kissed. All in all she could have been the queen of a campus much more densely populated with co-eds than Cavalier was. "You may call me Alis," she said. "And I'm nineteen." Don grinned. "Going on?" "Three months past. How old are you , Mr. Cort?" "Don's the name I've had for twenty-six years. Please use it." "Gladly. And now, Don, unless you want another cup of coffee, I'll go with you to the end of the world." "On such short notice?" Don was intrigued. Last night the redhead from the club car had repelled an advance that hadn't been made, and this morning a blonde was apparently making an advance that hadn't been solicited. He wondered where Geneva Jervis was, but only vaguely. "I'll admit to the double entendre ," Alis said. "What I meant—for now—was that we can stroll out to where Superior used to be attached to the rest of Ohio and see how the Earth is getting along without us." "Delighted. But don't you have any classes?" "Sure I do. Non-Einsteinian Relativity 1, at nine o'clock. But I'm a demon class-cutter, which is why I'm still a Senior at my advanced age. On to the brink!" They walked south from the campus and came to the railroad track. The train was standing there with nowhere to go. It had been abandoned except for the conductor, who had dutifully spent the night aboard. "What's happening?" he asked when he saw them. "Any word from down there?" "Not that I know of," Don said. He introduced him to Alis Garet. "What are you going to do?" "What can I do?" the conductor asked. "You can go over to Cavalier and have breakfast," Alis said. "Nobody's going to steal your old train." The conductor reckoned as how he might just do that, and did. "You know," Don said, "I was half-asleep last night but before the train stopped I thought it was running alongside a creek for a while." "South Creek," Alis said. "That's right. It's just over there." "Is it still? I mean hasn't it all poured off the edge by now? Was that Superior's water supply?" Alis shrugged. "All I know is you turn on the faucet and there's water. Let's go look at the creek." They found it coursing along between the banks. "Looks just about the same," she said. "That's funny. Come on; let's follow it to the edge." The brink, as Alis called it, looked even more awesome by daylight. Everything stopped short. There were the remnants of a cornfield, with the withered stalks cut down, then there was nothing. There was South Creek surging along, then nothing. In the distance a clump of trees, with a few autumn leaves still clinging to their branches, simply ended. "Where is the water going?" Don asked. "I can't make it out." "Down, I'd say. Rain for the Earth-people." "I should think it'd be all dried up by now. I'm going to have a look." "Don't! You'll fall off!" "I'll be careful." He walked cautiously toward the edge. Alis followed him, a few feet behind. He stopped a yard from the brink and waited for a spell of dizziness to pass. The Earth was spread out like a topographer's map, far below. Don took another wary step, then sat down. "Chicken," said Alis. She laughed uncertainly, then she sat down, too. "I still can't see where the water goes," Don said. He stretched out on his stomach and began to inch forward. "You stay there." Finally he had inched to a point where, by stretching out a hand, he could almost reach the edge. He gave another wriggle and the fingers of his right hand closed over the brink. For a moment he lay there, panting, head pressed to the ground. "How do you feel?" Alis asked. "Scared. When I get my courage back I'll pick up my head and look." Alis put a hand out tentatively, then purposefully took hold of his ankle and held it tight. "Just in case a high wind comes along," she said. "Thanks. It helps. Okay, here we go." He lifted his head. "Damn." "What?" "It still isn't clear. Do you have a pocket mirror?" "I have a compact." She took it out of her bag with her free hand and tossed it to him. It rolled and Don had to grab to keep it from going over the edge. Alis gave a little shriek. Don was momentarily unnerved and had to put his head back on the ground. "Sorry," she said. Don opened the compact and carefully transferred it to his right hand. He held it out beyond the edge and peered into it, focusing it on the end of the creek. "Now I've got it. The water isn't going off the edge!" "It isn't? Then where is it going?" "Down, of course, but it's as if it's going into a well, or a vertical tunnel, just short of the edge." "Why? How?" "I can't see too well, but that's my impression. Hold on now. I'm coming back." He inched away from the edge, then got up and brushed himself off. He returned her compact. "I guess you know where we go next." "The other end of the creek?" "Exactly." South Creek did not bisect Superior, as Don thought it might, but flowed in an arc through a southern segment of it. They had about two miles to go, past South Creek Bridge—which used to lead to Ladenburg, Alis said—past Raleigh Country Club (a long drive would really put the ball out of play, Don thought) and on to the edge again. But as they approached what they were forced to consider the source of the creek, they found a wire fence at the spot. "This is new," Alis said. The fence, which had a sign on it, warning—electrified , was semicircular, with each end at the edge and tarpaulins strung behind it so they could see the mouth of the creek. The water flowed from under the tarp and fence. "Look how it comes in spurts," Alis said. "As if it's being pumped." Smaller print on the sign said: Protecting mouth of South Creek, one of two sources of water for Superior. Electrical charge in fence is sufficient to kill. It was signed: Vincent Grande, Chief of Police, Hector Civek, Mayor . "What's the other source, besides the faucet in your bathroom?" Don asked. "North Lake, maybe," Alis said. "People fish there but nobody's allowed to swim." "Is the lake entirely within the town limits?" "I don't know." "If it were on the edge, and if I took a rowboat out on it, I wonder what would happen?" "I know one thing—I wouldn't be there holding your ankle while you found out." She took his arm as they gazed past the electrified fence at the Earth below and to the west. "It's impressive, isn't it?" she said. "I wonder if that's Indiana way over there?" He patted her hand absent-mindedly. "I wonder if it's west at all. I mean, how do we know Superior is maintaining the same position up here as it used to down there?" "We could tell by the sun, silly." "Of course," he said, grinning at his stupidity. "And I guess we're not high enough to see very far. If we were we'd be able to see the Great Lakes—or Lake Erie, anyway." They were musing about the geography when a plane came out of a cloudbank and, a second later, veered sharply. They could make out UAL on the underside of a wing. As it turned they imagined they could see faces peering out of the windows. They waved and thought they saw one or two people wave back. Then the plane climbed toward the east and was gone. "Well," Don said as they turned to go back to Cavalier, "now we know that they know. Maybe we'll begin to get some answers. Or, if not answers, then transportation." "Transportation?" Alis squeezed the arm she was holding. "Why? Don't you like it here?" "If you mean don't I like you, the answer is yes, of course I do. But if I don't get out of this handcuff soon so I can take a bath and get into clean clothes, you're not going to like me." "You're still quite acceptable, if a bit whiskery." She stopped, still holding his arm, and he turned so they were face to face. "So kiss me," she said, "before you deteriorate." They were in the midst of an extremely pleasant kiss when the brief case at the end of Don's handcuff began to talk to him. | A. He would have driven down into the pit where Superior was formerly located. |
What was Rolf looking for when he set off around the wall of the pit?
A. Garmon
B. Light
C. Food
D. Other survivors
| THE HAIRY ONES by BASIL WELLS Marooned on a world within a world, aided by a slim girl and an old warrior, Patrolman Sisko Rolf was fighting his greatest battle—to bring life to dying Mars. [Transcriber's Note: This etext was produced from Planet Stories Winter 1944. Extensive research did not uncover any evidence that the U.S. copyright on this publication was renewed.] "The outlaw ships are attacking!" Old Garmon Nash's harsh voice snapped like a thunderclap in the cramped rocket flyer's cabin. "Five or six of them. Cut the searchlights!" Sisko Rolf's stocky body was a blur of motion as he cut the rocket jets, doused the twin searchlights, and switched over to the audio beams that served so well on the surface when blind flying was in order. But here in the cavern world, thirty-seventh in the linked series of vast caves that underlie the waterless wastes of Mars, the reflected waves of sound were of little value. Distances were far too cramped—disaster might loom but a few hundred feet away. "Trapped us neatly," Rolf said through clenched teeth. "Tolled into their underground hideout by that water-runner we tried to capture. We can't escape, that's certain. They know these caverns better than.... We'll down some of them, though." "Right!" That was old Garmon Nash, his fellow patrolman aboard the Planet Patrol ship as he swung the deadly slimness of his rocket blast's barrel around to center on the fiery jets that betrayed the approaching outlaw flyers. Three times he fired the gun, the rocket projectiles blasting off with their invisible preliminary jets of gas, and three times an enemy craft flared up into an intolerable torch of flame before they realized the patrol ship had fired upon them. Then a barrage of enemy rocket shells exploded into life above and before them. Rolf swung the lax controls over hard as the bursts of fire revealed a looming barrier of stone dead ahead, and then he felt the tough skin of the flyer crumple inward. The cabin seemed to telescope about him. In a slow sort of wonder Rolf felt the scrape of rock against metal, and then the screeching of air through the myriad rents in the cabin's meralloy walls grew to a mad whining wail. Down plunged the battered ship, downward ever downward. Somehow Rolf found the strength to wrap his fingers around the control levers and snap on a quick burst from the landing rockets. Their mad speed checked momentarily, but the nose of the vertically plunging ship dissolved into an inferno of flame. The ship struck; split open like a rotten squash, and Rolf felt himself being flung far outward through thick blackness. For an eternity it seemed he hung in the darkness before something smashed the breath and feeling from his nerveless body. With a last glimmer of sanity he knew that he lay crushed against a rocky wall. Much later Rolf groaned with the pain of bruised muscles and tried to rise. To his amazement he could move all his limbs. Carefully he came to his knees and so to his feet. Not a bone was broken, unless the sharp breathlessness that strained at his chest meant cracked ribs. There was light in the narrow pit in which he found himself, light and heat from the yet-glowing debris of the rocket flyer. The outlaws had blasted the crashed ship, his practiced eyes told him, and Garmon Nash must have died in the wreckage. He was alone in the waterless trap of a deep crevice. In the fading glow of the super-heated metal the vertical walls above mocked him. There could be no ascent from this natural prison-pit, and even if there were he could never hope to reach the surface forty miles and more overhead. The floors of the thirty-seven caves through which they had so carefully jetted were a splintered, creviced series of canyon-like wastes, and as he ascended the rarefied atmosphere of the higher levels would spell death. Rolf laughed. Without a pressure mask on the surface of Mars an Earthman was licked. Without water and food certain death grinned in his face, for beyond the sand-buried entrance to these lost equatorial caves there were no pressure domes for hundreds of miles. Here at least the air was thick enough to support life, and somewhere nearby the outlaws who smuggled their precious contraband water into the water-starved domes of North Mars lay hidden. The young patrolman unzippered his jacket pocket and felt for the emergency concentrate bars that were standard equipment. Half of the oval bar he crushed between his teeth, and when the concentrated energy flooded into his muscles he set off around the irregular wall of the pit. He found the opening less than ten paces from the starting point, an empty cavity higher than a man and half as wide. The glow from the gutted ship was failing and he felt for the solar torch that hugged flatly against his hip. He uncapped the torch and the miniature sun glowed redly from its lensed prison to reveal the rocky corridor stretching out ahead. Light! How many hours later it was when the first faint glow of white light reached his eyes Rolf did not know—it had seemed an eternity of endless plodding along that smooth-floored descending tunnel. Rolf capped the solar torch. No use wasting the captive energy needlessly he reasoned. And he loosened the expoder in its holster as he moved carefully forward. The outlaw headquarters might be close ahead, headquarters where renegade Frogs, Venusians from the southern sunken marshes of Mars, and Earthmen from dusty North Mars, concealed their precious hoard of water from the thirsty colonists of North Mars. "They may have found the sunken seas of Mars," thought Rolf as he moved alertly forward, "water that would give the mining domes new life." His fists clenched dryly. "Water that should be free!" Then the light brightened before him as he rounded a shouldering wall of smoothly trimmed stone, and the floor fell away beneath his feet! He found himself shooting downward into a vast void that glowed softly with a mysterious all-pervading radiance. His eyes went searching out, out into undreamed distance. For miles below him there was nothing but emptiness, and for miles before him there was that same glowing vacancy. Above the cavern's roof soared majestically upward; he could see the narrow dark slit through which his feet had betrayed him, and he realized that he had fallen through the vaulted rocky dome of this fantastic abyss. It was then, even as he snapped the release of his spinner and the nested blades spun free overhead, that he saw the slowly turning bulk of the cloud-swathed world, a tiny five mile green ball of a planet! The weird globe was divided equally into hemispheres, and as the tiny world turned between its confining columns a green, lake-dotted half alternated with a blasted, splintered black waste of rocky desert. As the spinner dropped him slowly down into the vast emptiness of the great shining gulf, Rolf could see that a broad band of stone divided the green fertile plains and forests from the desolate desert wastes of the other half. Toward this barrier the spinner bore him, and Rolf was content to let it move in that direction—from the heights of the wall he could scout out the country beyond. The wall expanded as he came nearer to the pygmy planet. The spinner had slowed its speed; it seemed to Rolf that he must be falling free in space for a time, but the feeble gravity of the tiny world tugged at him more strongly as he neared the wall. And the barrier became a jumbled mass of roughly-dressed stone slabs, from whose earth-filled crevices sprouted green life. So slowly was the spinner dropping that the blackened desolation of the other hemisphere came sliding up beneath his boots. He looked down into great gashes in the blackness of the desert and saw there the green of sunken oases and watered canyons. He drifted slowly toward the opposite loom of the mysterious wall with a swift wind off the desert behind him. A hundred yards from the base of the rocky wall his feet scraped through black dust, and he came to a stop. Deftly Rolf nested the spinners again in their pack before he set out toward the heaped-up mass of stone blocks that was the wall. Ten steps he took before an excited voice called out shrilly from the rocks ahead. Rolf's slitted gray eyes narrowed yet more and his hand dropped to the compact expoder machine-gun holstered at his hip. There was the movement of a dark shape behind the screen of vines and ragged bushes. "Down, Altha," a deeper voice rumbled from above, "it's one of the Enemy." The voice had spoken in English! Rolf took a step forward eagerly and then doubt made his feet falter. There were Earthmen as well as Frogs among the outlaws. This mysterious world that floated above the cavern floor might be their headquarters. "But, Mark," the voice that was now unmistakably feminine argued, "he wears the uniform of a patrolman." "May be a trick." The deep voice was doubtful. "You know their leader, Cannon, wanted you. This may be a trick to join the Outcasts and kidnap you." The girl's voice was merry. "Come on Spider-legs," she said. Rolf found himself staring, open-mouthed, at the sleek-limbed vision that parted the bushes and came toward him. A beautiful woman she was, with the long burnished copper of her hair down around her waist, but beneath the meager shortness of the skin tunic he saw that her firm flesh was covered with a fine reddish coat of hair. Even her face was sleek and gleaming with its coppery covering of down. "Hello, patrol-a-man," she said shyly. An elongated pencil-ray of a man bounced nervously out to her side. "Altha," he scolded, scrubbing at his reddened bald skull with a long-fingered hand, "why do you never listen to me? I promised your father I'd look after you." He hitched at his tattered skin robe. The girl laughed, a low liquid sound that made Rolf's heart pump faster. "This Mark Tanner of mine," she explained to the patrolman, "is always afraid for me. He does not remember that I can see into the minds of others." She smiled again as Rolf's face slowly reddened. "Do not be ashamed," she said. "I am not angry that you think I am—well, not too unattractive." Rolf threw up the mental block that was the inheritance from his grueling years of training on Earth Base. His instructors there had known that a few gifted mortals possess the power of a limited telepathy, and the secrets of the Planet Patrol must be guarded. "That is better, perhaps." The girl's face was demure. "And now perhaps you will visit us in the safety of the vaults of ancient Aryk." "Sorry," said the tall man as Rolf sprang easily from the ground to their side. "I'm always forgetting the mind-reading abilities of the Hairy People." "She one of them?" Rolf's voice was low, but he saw Altha's lip twitch. "Mother was." Mark Tanner's voice was louder. "Father was Wayne Stark. Famous explorer you know. I was his assistant." "Sure." Rolf nodded. "Lost in equatorial wastelands—uh, about twenty years ago—2053, I believe." "Only we were not lost on the surface," explained Tanner, his booming voice much too powerful for his reedy body, "Wayne Stark was searching for the lost seas of Mars. Traced them underground. Found them too." He paused to look nervously out across the blasted wasteland. "We ran out of fuel here on Lomihi," he finished, "with the vanished surface waters of Mars less than four miles beneath us." Rolf followed the direction of the other's pale blue eyes. Overhead now hung the bottom of the cavern. An almost circular island of pale yellow lifted above the restless dark waters of a vast sea. Rolf realized with a wrench of sudden fear that they actually hung head downward like flies walking across a ceiling. "There," roared Tanner's voice, "is one of the seas of Mars." "One," repeated Rolf slowly. "You mean there are more?" "Dozens of them," the older man's voice throbbed with helpless rage. "Enough to make the face of Mars green again. Cavern after cavern lies beyond this first one, their floors flooded with water." Rolf felt new strength pump into his tired bruised muscles. Here lay the salvation of Earth's thirsting colonies almost within reach. Once he could lead the scientists of North Mars to this treasure trove of water.... "Mark!" The girl's voice was tense. Rolf felt her arm tug at his sleeve and he dropped beside her in the shelter of a clump of coarse-leaved gray bushes. "The Furry Women attack!" A hundred paces away Rolf made the dark shapes of armed warriors as they filed downward from the Barrier into the blackened desolation of the desert half of Lomihi. "Enemies?" he whispered to Mark Tanner hoarsely. "Right." The older man was slipping the stout bowstring into its notched recess on the upper end of his long bow. "They cross the Barrier from the fertile plains of Nyd to raid the Hairy People. They take them for slaves." "I must warn them." Altha's lips thinned and her brown-flecked eyes flamed. "The outlaws may capture," warned Tanner. "They have taken over the canyons of Gur and Norpar, remember." "I will take the glider." Altha was on her feet, her body crouched over to take advantage of the sheltering shrubs. She threaded her way swiftly back along a rocky corridor in the face of the Barrier toward the ruins of ancient Aryk. Tanner shrugged his shoulders. "What can I do? Altha has the blood of the Hairy People in her veins. She will warn them even though the outlaws have turned her people against her." Rolf watched the column of barbarically clad warriors file out upon the barren desert and swing to the right along the base of the Barrier. Spear tips and bared swords glinted dully. "They will pass within a few feet!" he hissed. "Right." Tanner's fingers bit into Rolf's arm. "Pray that the wind does not shift, their nostrils are sensitive as those of the weasels they resemble." Rolf's eyes slitted. There was something vaguely unhuman about those gracefully marching figures. He wondered what Tanner had meant by calling them weasels, wondered until they came closer. Then he knew. Above half naked feminine bodies, sinuous and supple as the undulating coils of a serpent, rose the snaky ditigrade head of a weasel-brute! Their necks were long and wide, merging into the gray-furred muscles of their narrow bodies until they seemed utterly shoulderless, and beneath their furry pelts the ripples of smooth-flowing muscles played rhythmically. There was a stench, a musky penetrating scent that made the flesh of his body crawl. "See!" Tanner's voice was muted. "Giffa, Queen of the Furry Ones!" Borne on a carved and polished litter of ebon-hued wood and yellowed bone lolled the hideous queen of that advancing horde. Gaunt of body she was, her scarred gray-furred hide hanging loose upon her breastless frame. One eye was gone but the other gleamed, black and beady, from her narrow earless skull. And the skulls of rodents and men alike linked together into ghastly festoons about her heavy, short-legged litter. Men bore the litter, eight broad-shouldered red-haired men whose arms had been cut off at the shoulders and whose naked backs bore the weals of countless lashes. Their bodies, like that of Altha, were covered with a silky coat of reddish hair. Rolf raised his expoder, red anger clouding his eyes as he saw these maimed beasts of burden, but the hand of Mark Tanner pressed down firmly across his arm. The older man shook his head. "Not yet," he said. "When Altha has warned the Hairy People we can cut off their retreat. After they have passed I will arouse the Outcasts who live here upon the Barrier. Though their blood is that of the two races mingled they hate the Furry Ones." A shadow passed over their hiding place. The Furry Amazons too saw the indistinct darkness and looked up. High overhead drifted the narrow winged shape of a glider, and the warrior women shrieked their hatred. Gone now was their chance for a surprise attack on the isolated canyons of the Hairy People. They halted, clustered about their leader. Giffa snarled quick orders at them, her chisel-teeth clicking savagely. The column swung out into the wasteland toward the nearest sunken valleys of the Hairy People. Rolf and Mark Tanner came to their feet. Abruptly, then, the wind veered. From behind the two Earthmen it came, bearing the scent of their bodies out to the sensitive nostrils of the beast-women. Again the column turned. They glimpsed the two men and a hideous scrawling battle-cry burst from their throats. Rolf's expoder rattled briefly like a high-speed sewing machine as he flicked its muzzle back and forth along the ranks of attacking Furry Ones. Dozens of the hideous weasel creatures fell as the needles of explosive blasted them but hundreds more were swarming over their fallen sisters. Mark Tanner's bow twanged again and again as he drove arrows at the bloodthirsty warrior women. But the Furry Ones ran fearlessly into that rain of death. The expoder hammered in Rolf's heavy fist. Tanner smashed an elbow into Rolf's side. "Retreat!" he gasped. The Furry Amazons swarmed up over the lower terraces of rocks, their snaky heads thrust forward and their swords slashing. The two Earthmen bounded up and backward to the next jumbled layer of giant blocks behind them, their powerful earthly muscles negating Lomihi's feeble gravity. Spears showered thick about them and then they dropped behind the sheltering bulk of a rough square boulder. "Now where?" Rolf snapped another burst of expoder needles at the furry attackers as he asked. "To the vaults beneath the Forbidden City," Mark Tanner cried. "None but the Outcasts and we two have entered the streets of deserted Aryk." The bald scientist slung his bow over his head and one shoulder and went bounding away along a shadowy crevice that plunged raggedly into the heart of the Barrier. Rolf blasted another spurt of explosive needles at the Furry Ones and followed. Darkness thickened as they penetrated into the maze of the Barrier's shattered heart. An unseen furry shape sprang upon Rolf's shoulders and as he sank to his knees he felt hot saliva drip like acid upon his neck. His fist sent the attacker's bulk smashing against the rocky floor before fangs or claws could rip at his tender flesh, and he heard a choked snarl that ended convulsively in silence. Bat-winged blobs of life dragged wet leathery hide across his face, and beneath his feet slimy wriggling things crushed into quivering pulp. Then there was faint light again, and the high-vaulted roof of a rock dungeon rose above him. Mark Tanner was peering out a slitted embrasure that overlooked the desolate land of the Hairy People. Tanner's finger pointed. "Altha!" Rolf saw the graceful wings of the glider riding the thermals back toward the Barrier. "She had warned the Hairy People, and now she returns." "The weasel heads won't follow us here?" asked Rolf. Tanner laughed. "Hardly. They fear the spirits of the Ancients too much for that. They believe the invisible powers will drink their souls." "Then how about telling me about this hanging world?" "Simply the whim of an ancient Martian ruler. As I have learned from the inscriptions and metal tablets here in Aryk he could not conquer all of Mars so he created a world that would be all his own." Rolf laughed. "Like the pleasure globes of the wealthy on Earth." "Right." Tanner kept his eyes on the enlarging winged shape of Altha's flyer as he spoke. "Later, when the nations of Mars began draining off the seas and hoarding them in their underground caverns, Lomihi became a fortress for the few thousand aristocrats and slaves who escaped the surface wars. "The Hairy People were the rulers," he went on, "and the Furry Ones were their slaves. In the revolt that eventually split Lomihi into two warring races this city, Aryk, was destroyed by a strange vegetable blight and the ancient knowledge was lost to both races." "But," Rolf frowned thoughtfully, "what keeps Lomihi from crashing into the island? Surely the two columns at either end cannot support it?" "The island is the answer," said Tanner. "Somehow it blocks the force of gravity—shields Lomihi from...." He caught his breath suddenly. "The outlaws!" he cried. "They're after Altha." Rolf caught a glimpse of a sleek rocket flyer diving upon Altha's frail wing. He saw the girl go gliding steeply down toward a ragged jumble of volcanic spurs and pits and disappear from view. He turned to see the old man pushing another crudely constructed glider toward the outer wall of the rock chamber. Tanner tugged at a silvery metal bar inset into the stone wall. A section of the wall swung slowly inward. Rolf sprang to his side. "Let me follow," he said. "I can fly a glider, and I have my expoder." The older man's eyes were hot. He jerked at Rolf's hands and then suddenly thought better of it. "You're right," he agreed. "Help her if you can. Your weapon is our only hope now." Rolf pushed up and outward with all the strength of his weary muscles. The glider knifed forward with that first swift impetus, and drove out over the Barrier. The Furry Ones were struggling insect shapes below him, and he saw with a thrill that larger bodied warriors, whose bodies glinted with a dull bronze, were attacking them from the burnt-out wastelands. The Hairy People had come to battle the invaders. He guided the frail wing toward the shattered badlands where the girl had taken shelter, noting as he did so that the rocket flyer had landed near its center in a narrow strip of rocky gulch. A sudden thought made him grin. He drove directly toward the grounded ship. With this rocket flyer he could escape from Lomihi, return through the thirty-seven caverns to the upper world, and give to thirsty Mars the gift of limitless water again. A man stood on guard just outside the flyer's oval door. Rolf lined up his expoder and his jaw tensed. He guided the tiny soarer closer with one hand. If he could crash the glider into the guard, well and good. There would be no explosion of expoder needles to warn the fellow's comrades. But if the outlaw saw him Rolf knew that he would be the first to fire—his was the element of surprise. A score of feet lay between them, and suddenly the outlaw whirled about. Rolf pressed the firing button; the expoder clicked over once and the trimmer key jammed, and the doughy-faced Venusian swung up his own long-barreled expoder! Rolf snapped his weapon overhand at the Frog's hairless skull. The fish-bellied alien ducked but his expoder swung off the target momentarily. In that instant Rolf launched himself from the open framework of the slowly diving glider, full upon the Venusian. They went down, Rolf swinging his fist like a hammer. He felt the Frog go limp and he loosed a relieved whistle. Now with a rocket flyer and the guard's rifle expoder in his grasp the problem of escape from the inner caverns was solved. He would rescue the girl, stop at the Forbidden City for Mark Tanner, and blast off for the upper crust forty miles and more overhead. He knelt over the prostrate Venusian, using his belt and a strip torn from his greenish tunic to bind the unconscious man. The knots were not too tight, the man could free himself in the course of a few hours. He shrugged his shoulders wearily and started to get up. A foot scraped on stone behind him. He spun on bent knees and flung himself fifty feet to the further side of the narrow gulch with the same movement. Expoder needles splintered the rocks about him as he dropped behind a sheltering rocky ledge, and he caught a glimpse of two green-clad men dragging the bronze-haired body of the girl he had come to save into the shelter of the flyer. A green bulge showed around the polished fuselage and Rolf pressed his captured weapon's firing button. A roar of pain came from the wounded man, and he saw an outflung arm upon the rocky ground that clenched tightly twice and relaxed to move no more. The outlaw weapon must have been loaded with a drum of poisoned needles, the expoder needles had not blasted a vital spot in the man's body. The odds were evening, he thought triumphantly. There might be another outlaw somewhere out there in the badlands, but no more than that. The flyer was built to accommodate no more than five passengers and four was the usual number. He shifted his expoder to cover the opposite end of the ship's squatty fuselage. And something that felt like a mountain smashed into his back. He was crushed downward, breathless, his eyes glimpsing briefly the soiled greenish trousers of his attacker as they locked on either side of his neck, and then blackness engulfed him as a mighty sledge battered endlessly at his skull. This sledge was hammering relentlessly as Rolf sensed his first glimmer of returning light. There were two sledges, one of them that he identified as the hammering of blood in his throbbing temples, and the other the measured blasting pulse of rocket jets. He opened his eyes slowly to find himself staring at the fine-crusted metal plates of a flyer's deck. His nose was grinding into the oily muck that only undisciplined men would have permitted to accumulate. Cautiously his head twisted until he could look forward toward the controls. The bound body of Altha Stark faced him, and he saw her lips twist into a brief smile of recognition. She shook her head and frowned as he moved his arm. But Rolf had learned that his limbs were not bound—apparently the outlaws had considered him out of the blasting for the moment. By degrees Rolf worked his arm down to his belt where his solar torch was hooked. His fingers made careful adjustments within the inset base of the torch, pushing a lever here and adjusting a tension screw there. The ship bumped gently as it landed and the thrum of rockets ceased. The cabin shifted with the weight of bodies moving from their seats. Rolf heard voices from a distance and the answering triumphant bawling of his two captors. The moment had come. He turned the cap of the solar torch away from his body and freed it. Heat blasted at his body as the stepped-up output of the torch made the oily floor flame. He lay unmoving while the thick smoke rolled over him. "Fire!" There was panic in the outlaw's voice. Rolf came to his knees in the blanketing fog and looked forward. One of the men flung himself out the door, but the other reached for the extinguisher close at hand. His thoughts were on the oily smoke; not on the prisoners, and so the impact of Rolf's horizontally propelled body drove the breath from his lungs before his hand could drop to his belted expoder. The outlaw was game. His fists slammed back at Rolf, and his knees jolted upward toward the patrolman's vulnerable middle. But Rolf bored in, his own knotted hands pumping, and his trained body weaving instinctively aside from the crippling blows aimed at his body. For a moment they fought, coughing and choking from the thickening pall of smoke, and then the fingers of the outlaw clamped around Rolf's throat and squeezed hard. The patrolman was weary; the wreck in the upper cavern and the long trek afterward through the dark tunnels had sapped his strength, and now he felt victory slipping from his grasp. He felt something soft bump against his legs, legs so far below that he could hardly realize that they were his, and then he was falling with the relentless fingers still about his throat. As from a great distant he heard a cry of pain and the blessed air gulped into his raw throat. His eyes cleared. He saw Altha's bound body and head. Her jaws were clamped upon the arm of the outlaw and even as he fought for more of the reeking smoky air of the cabin he saw the man's clenched fist batter at her face. Rolf swung, all the weight of his stocky body behind the blow, and the outlaw thudded limply against the opposite wall of the little cabin. No time to ask the girl if she were injured. The patrolman flung himself into the spongy control chair's cushions and sent the ship rocketing skyward. Behind him the thin film of surface oil no longer burned and the conditioning unit was clearing the air. "Patrolman," the girl's voice was beside him. "We're safe!" "Everything bongo?" Rolf wanted to know. "Of course," she smiled crookedly. "Glad of that." Rolf felt the warmth of her body so close beside him. A sudden strange restlessness came with the near contact. Altha smiled shyly and winced with pain. "Do you know," she said, "even yet I do not know your name." Rolf grinned up at her. "Need to?" he asked. The girl's eyes widened. A responsive spark blazed in them. "Handier than calling you Shorty all the time," she quipped. Then they were over the Barrier and Rolf saw the last of the beaten Furry Ones racing back across the great wall toward the Plains of Nyd. He nosed the captured ship down toward the ruined plaza of the Forbidden City. Once Mark Tanner was aboard they would blast surfaceward with their thrilling news that all Mars could have water in plenty again. Rolf snorted. "Shorty," he said disgustedly as they landed, but his arm went out toward the girl's red-haired slimness, and curved around it. | B. Light |
GP most likely stands for?
A. generic pharmaceutical
B. ghost publisher
C. geriatric patient
D. general practitioner
| AI: what's the worst that could happen? The Centre for the Future of Intelligence is seeking to investigate the implications of artificial intelligence for humanity, and make sure humans take advantage of the opportunities while dodging the risks. It launched at the University of Cambridge last October, and is a collaboration between four universities and colleges – Cambridge, Oxford, Imperial and Berkeley – backed with a 10-year, £10m grant from the Leverhulme Trust. Because no single discipline is ideally suited to this task, the centre emphasises the importance of interdisciplinary knowledge-sharing and collaboration. It is bringing together a diverse community of some of the world's best researchers, philosophers, psychologists, lawyers and computer scientists. Executive director of the centre is Stephen Cave, a writer, philosopher and former diplomat. Harry Armstrong, head of futures at Nesta, which publishes The Long + Short, spoke with Cave about the impact of AI. Their conversation has been edited. Harry Armstrong: Do you see the interdisciplinary nature of the centre as one of its key values and one of the key impacts you hope it will have on the field? Stephen Cave: Thinking about the impact of AI is not something that any one discipline owns or does in any very systematic way. So if academia is going to rise to the challenge and provide thought leadership on this hugely important issue, then we’re going to need to do it by breaking down current disciplinary boundaries and bringing people with very different expertise together. That means bringing together the technologists and the experts at developing these algorithms together with social scientists, philosophers, legal scholars and so forth. I think there are many areas of science where more interdisciplinary engagement would be valuable. Biotech’s another example. In that sense AI isn’t unique, but I think because thinking about AI is still in very early stages, we have an opportunity to shape the way in which we think about it, and build that community. We want to create a space where many different disciplines can come together and develop a shared language, learn from each other’s approaches, and hopefully very quickly move to be able to actually develop new ideas, new conclusions, together. But the first step is learning how to talk to each other. At a recent talk, Naomi Klein said that addressing the challenge of climate change could not have come at a worse time. The current dominant political and economic ideologies, along with growing isolationist sentiment, runs contrary to the bipartisan, collaborative approaches needed to solve global issues like climate change. Do you see the same issues hampering a global effort to respond to the challenges AI raises? Climate change suffers from the problem that the costs are not incurred in any direct way by the industrialists who own the technology and are profiting from it. With AI, that has been the case so far; although not on the same scale. There has been disruption but so far, compared to industrialisation, the impact has been fairly small. That will probably change. AI companies, and in particular the big tech companies, are very concerned that this won't go like climate change, but rather it will go like GMOs: that people will have a gut reaction to this technology as soon as the first great swathe of job losses take hold. People speculate that 50m jobs could be lost in the US if trucking is automated, which is conceivable within 10 years. You could imagine a populist US government therefore simply banning driverless cars. So I think there is anxiety in the tech industry that there could be a serious reaction against this technology at any point. And so my impression is that there is a feeling within these companies that these ethical and social implications need to be taken very seriously, now. And that a broad buy-in by society into some kind of vision of the future in which this technology plays a role is required, if a dangerous – or to them dangerous – counteraction is to be avoided. My personal experience working with these tech companies is that they are concerned for their businesses and genuinely want to do the right thing. Of course there are intellectual challenges and there is money to be made, but equally they are people who don't think when they get up in the morning that they're going to put people out of jobs or bring about the downfall of humanity. As the industry matures it's developing a sense of responsibility. So I think we've got a real opportunity, despite the general climate, and in some ways because of it. There's a great opportunity to bring industry on board to make sure the technology is developed in the right way. One of the dominant narratives around not only AI but technology and automation more generally is that we, as humans, are at the mercy of technological progress. If you try and push against this idea you can be labelled as being anti-progress and stuck in the past. But we do have a lot more control than we give ourselves credit for. For example, routineness and susceptibility to automation are not inevitable features of occupations, job design is hugely important. How do we design jobs? How do we create jobs that allow people to do the kind of work they want to do? There can be a bit of a conflict between being impacted by what's happening and having some sort of control over what we want to happen. Certainly, we encounter technological determinism a lot. And it's understandable. For us as individuals, of course it does feel like it always is happening and we just have to cope. No one individual can do much about it, other than adapt. But that's different when we consider ourselves at a level of a society, as a polis [city state], or as an international community. I think we can shape the way in which technology develops. We have various tools. In any given country, we have regulations. There's a possibility of international regulation. Technology is emerging from a certain legal, political, normative, cultural, and social framework. It's coming from a certain place. And it is shaped by all of those things. And I think the more we understand a technology's relationship with those things, and the more we then consciously try to shape those things, the more we are going to influence the technology. So, for example, developing a culture of responsible innovation. For example, a kind of Hippocratic oath for AI developers. These things are within the realms of what is feasible, and I think will help to shape the future. One of the problems with intervention, generally, is that we cannot control the course of events. We can attempt to, but we don't know how things are going to evolve. The reality is, societies are much too complex for us to be able to shape them in any very specific way, as plenty of ideologies and political movements have found to their cost. There are often unforeseen consequences that can derail a project. I think, nonetheless, there are things we can do. We can try to imagine how things might go very badly wrong, and then work hard to develop systems that will stop that from happening. We can also try collectively to imagine how things could go very right. The kind of society that we actually want to live in that uses this technology. And I'm sure that will be skewed in all sorts of ways, and we might imagine things that seem wonderful and actually have terrible by-products. This conversation cannot be in the hands of any one group. It oughtn't be in the hands of Silicon Valley billionaires alone. They've got their role to play, but this is a conversation we need to be having as widely as possible. The centre is developing some really interesting projects but perhaps one of the most interesting is the discussion of what intelligence might be. Could you go into a bit more detail about the kinds of questions you are trying to explore in this area? You mean kinds of intelligence? Yeah. I think this is very important because historically, we've had an overwhelming tendency to anthropomorphise. We define what intelligence is, historically, as being human-like. And then within that, being like certain humans. And it's taken a very long time for the academic community to accept that there could be such a thing as non-human intelligence at all. We know that crows, for example, who have had a completely different evolutionary history, or octopuses, who have an even more different evolutionary history, might have a kind of intelligence that's very different to ours. That in some ways rivals our own, and so forth. But luckily, we have got to that point in recent years of accepting that we are not the only form of intelligence. But now, AI is challenging that from a different direction. Just as we are accepting that the natural world offers this enormous range of different intelligences, we are at the same time inventing new intelligences that are radically different to humans. And I think, still, this anthropomorphic picture of the kind of humanoid android, the robot, dominates our idea of what AI is too much. And too many people, and the industry as well, talk about human-level artificial intelligence as a goal, or general AI, which basically means like a human. But actually what we're building is nothing like a human. When the first pocket calculator was made, it didn't do maths like a human. It was vastly better. It didn't make the occasional mistake. When we set about creating these artificial agents to solve these problems, because they have a completely different evolutionary history to humans, they solve problems in very different ways. And until now, people have been fairly shy about describing them as intelligent. Or rather, in the history of AIs, we think solving a particular problem would require intelligence. Then we solve it. And then that's no longer intelligence, because we've solved it. Chess is a good example. But the reality is, we are creating a whole new world of different artificial agents. And we need to understand that world. We need to understand all the different ways of being clever, if you like. How you can be extremely sophisticated at some particular rational process, and yet extremely bad at another one in a way that bears no relation to the way humans are on these axes. And this is important, partly because we need to expand our sense of what is intelligent, like we have done with the natural world. Because lots of things follow from saying something is intelligent. Historically, we have a long tradition in Western philosophy of saying those who are intelligent should rule. So if intelligence equates to power, then obviously we need to think about what we mean by intelligence. Who has it and who doesn't. Or how it equates to rights and responsibilities. It certainly is a very ambitious project to create the atlas of intelligence. There was a point I read in something you wrote on our ideas of intelligence that I thought was very interesting. We actually tend to think of intelligence at the societal level when we think about human ability, rather than at the individual level but in the end conflate the two. I think that's a very good point, when we think about our capabilities, we think about what we can achieve as a whole, not individually. But when we talk about AI, we tend to think about that individual piece of technology, or that individual system. So for example if we think about the internet of things and AI, we should discuss intelligence as something encompassed by the whole. Yeah, absolutely. Yes, right now, perhaps it is a product of our anthropomorphising bias. But there is a tendency to see a narrative of AI versus humanity, as if it's one or the other. And yet, obviously, there are risks in this technology long before it acquires any kind of manipulative agency. Robotic technology is dangerous. Or potentially dangerous. But at the same time, most of what we're using technology for is to enhance ourselves, to increase our capacities. And a lot of what AI is going to be doing is augmenting us – we're going to be working as teams, AI-human teams. Where do you think this AI-human conflict, or concept of a conflict, comes from? Do you think that's just a reflection of historical conversations we've had about automation, or do you think it is a deeper fear? I do think it comes both from some biases that might well be innate, such as anthropomorphism, or our human tendency to ascribe agency to other objects, particularly moving ones, is well-established and probably has sound evolutionary roots. If it moves, it's probably wise to start asking yourself questions like, "What is it? What might it want? Where might it be going? Might it be hungry? Do I look like food to it?" I think it makes sense, it's natural for us to think in terms of agency. And when we do, it's natural for us to project our own ways of being and acting. And we, as primates, are profoundly co-operative. But at the same time, we're competitive and murderous. We have a strong sense of in-group versus out-group, which is responsible for both a great deal of cooperation, within the in-group, but also terrible crimes. Murder, rape, pillage, genocide; and they're pointed at the out-group. And so I think it's very natural for us to see AIs in terms of agents. We anthropomorphise them as these kind of android robots. And then we think about, well, you know, are they part of our in-group, or are they some other group? If they're some other group, it's us against them. Who's going to win? Well, let's see. So I think that's very natural, I think that's very human. There is this long tradition, in Western culture in particular, with associating intelligence and dominance and power. It's interesting to speculate about how, and I wish I knew more about it, and I'd like to see more research on this, about how different cultures perceive AI. It's well known that Japan is very accepting of technology and robots, for example. You can think, well, we in the West have long been justifying power relations of a certain kind on the basis that we're 'cleverer'. That's why men get to vote and women don't, or whatever. In a culture where power is not based on intelligence but, say, on a caste system, which is purely hereditary, we’d build an AI, and it would just tune in, drop out, attain enlightenment, just sit in the corner. Or we beg it to come back and help us find enlightenment. It might be that we find a completely different narrative to the one that's dominant in the West. One of the projects the centre is running is looking into what kind of AI breakthroughs may come, when and what the social consequences could be. What do you think the future holds? What are your fears – what do you think could go right and wrong in the short, medium and long term? That's a big question. Certainly I don't lie awake at night worried that robots are going to knock the door down and come in with a machine gun. If the robots take over the world, it won't be by knocking the door down. At the moment, I think it's certainly as big a risk that we have a GMO moment, and there's a powerful reaction against the technology which prevents us from reaping the benefits, which are enormous. I think that's as big a risk as the risks from the technologies themselves. I think one worry that we haven't talked about is that we've become extremely dependent upon this technology. And that we essentially become deskilled. There's an extent to which the history of civilisation is the history of the domestication of the human species sort of by ourselves, and also by our technology, to some extent. And AI certainly allows for that to reach a whole new level. Just think about GPs with diagnostic tools. Even now, my GP consults the computer fairly regularly. But as diagnostic tools get better, what are they going to be doing other than just typing something into the computer and reading out what comes back? At which point, you might as well do away with the GP. But then, who does know about medicine? And so we do need to worry about deskilling and about becoming dependent. And it is entirely possible that you can imagine a society in which we're all sort of prosperous, in a sense. Our basic bodily needs are provided for, perhaps, in a way, to an extent that we've never before even dreamed of. Unprecedented in human history. And yet, we're stripped of any kind of meaningful work. We have no purpose. We're escaping to virtual reality. And then you could imagine all sorts of worrying countercultures or Luddite movements or what have you. I guess that's the kind of scenario that – I haven't sketched it terribly well – but that's the kind of thing that worries me more than missile-toting giant robots. As to utopian, yes, that's interesting. I certainly mentioned a couple of things. One thing that I hope is that this new technological revolution enables us to undo some of the damage of the last one. That's a very utopian thought and not terribly realistic, but we use fossil fuels so incredibly efficiently. The idea that driverless cars that are shared, basically a kind of shared service located off a Brownfield site does away with 95 per cent of all cars, freeing up a huge amount of space in the city to be greener, many fewer cars need to be produced, they would be on the road much less, there'd be fewer traffic jams. It's just one example, but the idea that we can live much more resource-efficiently, because we are living more intelligently through using these tools. And therefore can undo some of the damage of the last Industrial Revolution. That's my main utopian hope, I guess. Vintage toy robot image by josefkubes/Shutterstock This article was originally published on TheLong+Short. Read the original article. | D. general practitioner |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.